Pranava Madhyastha (@foobarin) 's Twitter Profile
Pranava Madhyastha

@foobarin

NLP and Multimodal Machine Learning. Asst. Professor @CityUniLondon

ID: 3556130595

linkhttps://pmadhyastha.github.io/ calendar_today05-09-2015 14:15:55

391 Tweet

147 Followers

986 Following

Armen Aghajanyan (@armenagha) 's Twitter Profile Photo

One core learning we had with Chameleon is that the intended form of the modality is a modality in itself. Visual Perception and Visual Generation are two separate modalities and must be treated as such; hence, using discretized tokens for perception is wrong.

One core learning we had with Chameleon is that the intended form of the modality is a modality in itself. Visual Perception and Visual Generation are two separate modalities and must be treated as such; hence, using discretized tokens for perception is wrong.
Bartłomiej Cupiał (@cupiabart) 's Twitter Profile Photo

So here's a story of, by far, the weirdest bug I've encountered in my CS career. Along with Maciej Wołczyk we've been training a neural network that learns how to play NetHack, an old roguelike game, that looks like in the screenshot. Recenlty, something unexpected happened.

So here's a story of, by far, the weirdest bug I've encountered in my CS career.

Along with <a href="/maciejwolczyk/">Maciej Wołczyk</a> we've been training a neural network that learns how to play NetHack, an old roguelike game, that looks like in the screenshot. Recenlty, something unexpected happened.
Dr Christopher Madan 🐘🧠💻 (he/him) (@cmadan) 's Twitter Profile Photo

"A random half of panelists were shown a CV and only a one-paragraph summary of the proposed research, while the other half were shown a CV and a full proposal. We find that withholding proposal texts from panelists did not detectibly impact rankings." link.springer.com/article/10.100…

Arvind Narayanan (@random_walker) 's Twitter Profile Photo

New essay: ML seems to promise discovery without understanding, but this is fool's gold that has led to a reproducibility crisis in ML-based science. aisnakeoil.com/p/scientists-s… (with Sayash Kapoor). In 2021 we compiled evidence that an error called leakage is pervasive in ML models

New essay: ML seems to promise discovery without understanding, but this is fool's gold that has led to a reproducibility crisis in ML-based science. aisnakeoil.com/p/scientists-s… (with <a href="/sayashk/">Sayash Kapoor</a>).

In 2021 we compiled evidence that an error called leakage is pervasive in ML models
Christopher Manning (@chrmanning) 's Twitter Profile Photo

I agree with much of both @emilymbender.bsky.social’s #ACL2024 presidential talk and (((ل()(ل() 'yoav))))👾’s rejoinder, but I want to comment on just one aspect where I disagree with both: the definition and domain of CL vs NLP. 🧵👇

Tal Linzen (@tallinzen) 's Twitter Profile Photo

as SAC for EMNLP I was asked to read the discussions between the authors and reviewers and had every intention to do so but the length of the discussions is out of control. many tables with results of new experiments, hundreds of lines of code (!). bring back word limits please.

Delip Rao e/σ (@deliprao) 's Twitter Profile Photo

Unless you are an OpenAI employee working on improving their products, I don’t understand why such efforts are science. Why are we (question to faculty) spending taxpayer dollars in doing QA for a closed product by a well-capitalized company that does not give back to science?

Unless you are an OpenAI employee working on improving their products, I don’t understand why such efforts are science. Why are we (question to faculty) spending taxpayer dollars in doing QA for a closed product by a well-capitalized company that does not give back to science?
kyutai (@kyutai_labs) 's Twitter Profile Photo

Today, we release several Moshi artifacts: a long technical report with all the details behind our model, weights for Moshi and its Mimi codec, along with streaming inference code in Pytorch, Rust and MLX. More details below 🧵 ⬇️ Paper: kyutai.org/Moshi.pdf Repo:

Emile van Krieken (@emilevankrieken) 's Twitter Profile Photo

Great workshop at AAAI about low-rank representations! These have important consequences for Neurosymbolic: Logical circuits can be understood as low-rank factorisations.

Edward Grefenstette (@egrefen) 's Twitter Profile Photo

This “English(/French/etc) is the new code” trope bemuses me, even as someone bullish about LLMs substantially changing how we code. A 🧵(1/9)

antonio vergari - hiring PhD students (@tetraduzione) 's Twitter Profile Photo

great opportunity to do a #PhD in #Europe in #ML #AI 🚨🚨🚨 I'll hire 2 students via #ELLIS this year, reach out if you want to do research in: - #reliable and #efficient ML in the wild - scalable #neurosymbolic #nesy AI - #lowrank representations - #tractable inference 🚨🚨🚨

(((ل()(ل() 'yoav))))👾 (@yoavgo) 's Twitter Profile Photo

i think this is the wrong question. yes, CS graduates are very bad in software development, and a dedicated LLM can be better. but put these graduates in a job, and some will develop to be "senior devs" at some point, capable of working on real-life, large systems. LLMs won't.

(((ل()(ل() 'yoav))))👾 (@yoavgo) 's Twitter Profile Photo

the term "experts" in "mixture of experts" in the context of LLMs is highly misleading and does way more harm than good in coming up with a conceptual representation of what this component brings to the table.

Christopher Manning (@chrmanning) 's Twitter Profile Photo

Papers at #EMNLP2024 #3 A counter-example to the frequently adopted mech interp linear representation hypothesis: Recurrent Neural Networks Learn … Non-Linear Representations Fri Nov 15 BlackboxNLP 2024 poster aclanthology.org/2024.blackboxn… CC Csordás Róbert Christopher Potts

Papers at #EMNLP2024 #3

A counter-example to the frequently adopted mech interp linear representation hypothesis: Recurrent Neural Networks Learn … Non-Linear Representations

Fri Nov 15 BlackboxNLP 2024 poster

aclanthology.org/2024.blackboxn…

CC <a href="/robert_csordas/">Csordás Róbert</a> <a href="/ChrisGPotts/">Christopher Potts</a>
(((ل()(ل() 'yoav))))👾 (@yoavgo) 's Twitter Profile Photo

(a) how did MMLU become the defacto standard benchmark every LLM is trying to beat? (b) it is estimated to contain 9% questions that human experts think are wrong. do we know if humans and models agree on which ones belong in this 9%?