Usha Bhalla (@ushabhalla_) 's Twitter Profile
Usha Bhalla

@ushabhalla_

ML Intepretability, Harvard CS PhD Candidate, Kempner Graduate Fellow (she/they)

ID: 832743085071142912

calendar_today18-02-2017 00:06:37

12 Tweet

92 Followers

139 Following

ellie schnitt! (@holy_schnitt) 's Twitter Profile Photo

me constantly forgetting to text back is a huge flaw but unfortunately there is literally nothing I can do about it! nothing at all case closed no further comments

π™·πš’πš–πšŠ π™»πšŠπš”πš”πšŠπš›πšŠπš“πšž (@hima_lakkaraju) 's Twitter Profile Photo

One of the biggest criticisms of the field of post hoc #XAI is that each method "does its own thing", it is unclear how these methods relate to each other & which methods are effective under what conditions. Our #NeurIPS2022 paper provides (some) answers to these questions. [1/N]

One of the biggest criticisms of the field of post hoc #XAI is that each method "does its own thing", it is unclear how these methods relate to each other & which methods are effective under what conditions. Our #NeurIPS2022 paper provides (some) answers to these questions. [1/N]
Ocean Vuong bot (@oceanvbot) 's Twitter Profile Photo

You’re not a monster, I said. But I lied. What I really wanted to say was that a monster is not such a terrible thing to be.

Alex Oesterling @ NeurIPS 2024 (@alex_oesterling) 's Twitter Profile Photo

Finally, I am pleased to announce πŸͺ’Interpreting CLIP with Sparse Linear Concept Embeddings (SpLiCE)πŸͺ’ Joint work with Usha Bhalla, as well as Suraj Srinivas, Flavio Calmon, and π™·πš’πš–πšŠ π™»πšŠπš”πš”πšŠπš›πšŠπš“πšž, which was just accepted to NeurIPS 2024! Check out the paper here: arxiv.org/abs/2402.10376

Finally, I am pleased to announce

πŸͺ’Interpreting CLIP with Sparse Linear Concept Embeddings (SpLiCE)πŸͺ’

Joint work with Usha Bhalla, as well as <a href="/Suuraj/">Suraj Srinivas</a>, <a href="/FlavioCalmon/">Flavio Calmon</a>, and <a href="/hima_lakkaraju/">π™·πš’πš–πšŠ π™»πšŠπš”πš”πšŠπš›πšŠπš“πšž</a>, which was just accepted to NeurIPS 2024! Check out the paper here:
arxiv.org/abs/2402.10376
π™·πš’πš–πšŠ π™»πšŠπš”πš”πšŠπš›πšŠπš“πšž (@hima_lakkaraju) 's Twitter Profile Photo

Super excited to share our latest preprint that unifies multiple areas within explainable AI that have been evolving somewhat independently: 1. Feature Attribution 2. Data Attribution 3. Model Component Attribution (aka Mechanistic Interpretability) arxiv.org/abs/2501.18887

Super excited to share our latest preprint that unifies multiple areas within explainable AI that have been evolving somewhat independently: 

1. Feature Attribution
2. Data Attribution
3. Model Component Attribution (aka Mechanistic Interpretability) 

arxiv.org/abs/2501.18887
Alex Oesterling @ NeurIPS 2024 (@alex_oesterling) 's Twitter Profile Photo

β€ΌοΈπŸ•šNew paper alert with Usha Bhalla: Leveraging the Sequential Nature of Language for Interpretability (openreview.net/pdf?id=hgPf1ki…)! 1/n