Dong Won (Don) Lee (@_dongwonlee) 's Twitter Profile
Dong Won (Don) Lee

@_dongwonlee

PhD at Personal Robots Group @MIT advised by @cynthiabreazeal, @lpmorency working on multimodal social AI agents. currently @MSFTResearch, prev. MS/BS @mldcmu

ID: 1465969247197319170

linkhttp://dongwonl.com calendar_today01-12-2021 09:01:30

117 Tweet

110 Followers

143 Following

Diyi Yang (@diyi_yang) 's Twitter Profile Photo

Learning social skills is out of reach for most people🙁 How can we make social skill training more accessible? We introduce 🌟APAM🌟 (AI Partner and AI Mentor) that leverages LLMs for social skill training via realistic practice and tailored feedback!

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments The first-of-its-kind scalable, real computer environment for multimodal agents, supporting task setup, execution-based evaluation, and interactive learning across various operating

Maria Khalusova (@mariakhalusova) 's Twitter Profile Photo

I put together a quick colab notebook using Llama-3-8B-Instruct for chatting with a PDF. Unstructured API for partitioning and chunking a large PDF file, FAISS for vector storage, LangChain for RAG, and quantized Llama-3-8B-Instruct (so that it fits on free Colab's GPU).

Rafael Rafailov @ NeurIPS (@rm_rafailov) 's Twitter Profile Photo

We have a new preprint out - your language model is not a reward, it’s a Q function! 1. The likelihood of the preferred answer must go down - it’s a policy divergence 2. MCTS guided decoding on language is equivalent to likelihood search on DPO 3. DPO learns credit assignment

We have a new preprint out - your language model is not a reward, it’s a Q function!
1. The likelihood of the preferred answer must go down - it’s a policy divergence
2. MCTS guided decoding on language is equivalent to likelihood search on DPO
3. DPO learns credit assignment
Xingyu Fu (@xingyufu2) 's Twitter Profile Photo

Can GPT-4V and Gemini-Pro perceive the world the way humans do? 🤔 Can they solve the vision tasks that humans can in the blink of an eye? 😉 tldr; NO, they are far worse than us 💁🏻‍♀️ Introducing BLINK👁 zeyofu.github.io/blink/, a novel benchmark that studies visual perception

Can GPT-4V and Gemini-Pro perceive the world the way humans do? 🤔

Can they solve the vision tasks that humans can in the blink of an eye? 😉

tldr; NO, they are far worse than us 💁🏻‍♀️

Introducing BLINK👁 zeyofu.github.io/blink/, a novel benchmark that studies visual perception
Leena Mathur (@lmathur_) 's Twitter Profile Photo

Curious about socially-intelligent AI? Check out our paper on underlying technical challenges, open questions, and opportunities to advance social intelligence in AI agents: Work w/ LP Morency, Paul Liang 📰Paper: arxiv.org/abs/2404.11023 💻Repo: github.com/l-mathur/socia… 🧵1/9

Curious about socially-intelligent AI? Check out our paper on underlying technical challenges, open questions, and opportunities to advance social intelligence in AI agents:

Work w/ <a href="/lpmorency/">LP Morency</a>, <a href="/pliang279/">Paul Liang</a> 

📰Paper: arxiv.org/abs/2404.11023
💻Repo: github.com/l-mathur/socia…

🧵1/9
andrew gao (@itsandrewgao) 's Twitter Profile Photo

Someone just dropped a dataset of 15 trillion tokens (as many as were used to train Llama 3)!!! Download this now before it gets taken down for “copyright reasons” Breakdown in thread 🧵 👇👇

Someone just dropped a dataset of 15 trillion tokens (as many as were used to train Llama 3)!!!

Download this now before it gets taken down for “copyright reasons”

Breakdown in thread 🧵 👇👇
Carlos E. Perez (@intuitmachine) 's Twitter Profile Photo

1/n A Ontology for Agentic AI Single Agent Architectures planning, self-correction, and suitability for straightforward tasks. Examples: * ReAct (Reason + Act): Iterative process of thought, action, and observation. * RAISE: ReAct with memory mechanism (short-term and

Carlos E. Perez (@intuitmachine) 's Twitter Profile Photo

AlphaLLM: An LLM that Learns and Improves Itself Large Language Models (LLMs) have revolutionized the field of Natural Language Processing, demonstrating remarkable capabilities in various tasks. However, they still struggle with complex reasoning and planning, often requiring

AlphaLLM: An LLM that Learns and Improves Itself

Large Language Models (LLMs) have revolutionized the field of Natural Language Processing, demonstrating remarkable capabilities in various tasks. However, they still struggle with complex reasoning and planning, often requiring
Abhinav Rao (@aethersura) 's Twitter Profile Photo

New paper on LLMs+culture! 🎊🎉 Thrilled to share our work on NormAd, a dataset evaluating whether LLMs can adapt to the diversity of cultural norms worldwide! (Spoiler: they can't!) ArXiv: arxiv.org/abs/2404.12464 w/ Akhila Yerukola Vishwa Shah Katharina Reinecke Maarten Sap (he/him) [1/n]

Yubin Kim (@ybkim95_ai) 's Twitter Profile Photo

I'm excited to share my recent publication in CHIL 2024, "Health-LLM: Large Language Models for Health Prediction via Wearable Sensor Data". Our study reveals the potential of LLMs as personal health learners with wearable sensors. Arxiv: arxiv.org/pdf/2401.06866…

I'm excited to share my recent publication in CHIL 2024, "Health-LLM: Large Language Models for Health Prediction via Wearable Sensor Data".

Our study reveals the potential of LLMs as personal health learners with wearable sensors.

Arxiv: arxiv.org/pdf/2401.06866…
Yi Tay (@yitayml) 's Twitter Profile Photo

New paper from Reka 🔥 (yes an actual paper). This time we're releasing part of our internal evals which we call Vibe-Eval 😃 This comprises of a hard set which imo is pretty challenging for frontier models today. The fun part here is that we constructed it by trying to

New paper from <a href="/RekaAILabs/">Reka</a> 🔥 (yes an actual paper).

This time we're releasing part of our internal evals which we call Vibe-Eval 😃 This comprises of a hard set which imo is pretty challenging for frontier models today. 

The fun part here is that we constructed it by trying to
Chief AI Officer (@chiefaioffice) 's Twitter Profile Photo

AI phone agents are here. Here are 8 startups enabling this you should know about + funding: 1. Bland AI - undisclosed - available as an API - $0.09/minute

Phillip Isola (@phillip_isola) 's Twitter Profile Photo

New paper: The Platonic Representation Hypothesis In which we posit that _different_ foundation models are converging to the _same_ representation of reality. paper: arxiv.org/abs/2405.07987 website: phillipi.github.io/prh/ code: github.com/minyoungg/plat… 1/8

Dan Hendrycks (@danhendrycks) 's Twitter Profile Photo

As an alternative to RLHF and adversarial training, we released short-circuiting. It makes models ~100x more robust. It works for LLMs, multimodal models, and agents. Unlike before, I now think robustly stopping models from generating harmful outputs may be highly tractable and

As an alternative to RLHF and adversarial training, we released short-circuiting.
It makes models ~100x more robust. It works for LLMs, multimodal models, and agents.

Unlike before, I now think robustly stopping models from generating harmful outputs may be highly tractable and
Paul Liang (@pliang279) 's Twitter Profile Photo

heading to #emnlp2024! would love to chat with those interested in joining our Multisensory Intelligence research group at MIT MIT Media Lab MIT EECS media.mit.edu/groups/multise… Our group studies the foundations of multisensory AI to create human-AI symbiosis across scales and sensory

heading to #emnlp2024! would love to chat with those interested in joining our Multisensory Intelligence research group at MIT <a href="/medialab/">MIT Media Lab</a> <a href="/MITEECS/">MIT EECS</a> 

media.mit.edu/groups/multise…

Our group studies the foundations of multisensory AI to create human-AI symbiosis across scales and sensory
Sherjil Ozair (@sherjilozair) 's Twitter Profile Photo

Very happy to hear that GANs are getting the test of time award at NeurIPS 2024. The NeurIPS test of time awards are given to papers which have stood the test of the time for a decade. I took some time to reminisce how GANs came about and how AI has evolve in the last decade.

AI at Meta (@aiatmeta) 's Twitter Profile Photo

🚀New from Meta FAIR: today we’re introducing Seamless Interaction, a research project dedicated to modeling interpersonal dynamics. The project features a family of audiovisual behavioral models, developed in collaboration with Meta’s Codec Avatars lab + Core AI lab, that