Gargi Ghosh (@gargighosh) 's Twitter Profile
Gargi Ghosh

@gargighosh

Researcher at FAIR (Meta AI)

ID: 95091994

linkhttps://scholar.google.com/citations?user=k5akwCcAAAAJ&hl=en&oi=ao calendar_today06-12-2009 23:32:46

66 Tweet

700 Followers

151 Following

Yannic Kilcher ๐Ÿ‡ธ๐Ÿ‡จ (@ykilcher) 's Twitter Profile Photo

๐Ÿ”ฅNew Video๐Ÿ”ฅ I delve (ha!) into Byte Latent Transformer: Patches Scale Better Than Tokens where the authors do away with tokenization and create an LLM architecture that operates on dynamically sized "patches" instead of tokens. By controlling the patch size, they gain a level

๐Ÿ”ฅNew Video๐Ÿ”ฅ
I delve (ha!) into Byte Latent Transformer: Patches Scale Better Than Tokens where the authors do away with tokenization and create an LLM architecture that operates on dynamically sized "patches" instead of tokens. By controlling the patch size, they gain a level
wh (@nrehiew_) 's Twitter Profile Photo

Wrote about some of my favourite papers over the past year or so and some research directions that I am excited about in 2025 As a bonus, I think it's a good overview for someone to catch up on the current state of the art :)

Wrote about some of my favourite papers over the past year or so and some research directions that I am excited about in 2025

As a bonus, I think it's a good overview for someone to catch up on the current state of the art :)
AI at Meta (@aiatmeta) 's Twitter Profile Photo

New research from Meta FAIR โ€” Meta Memory Layers at Scale. This work takes memory layers beyond proof-of-concept, proving their utility at contemporary scale โžก๏ธ go.fb.me/3lbt4m

clem ๐Ÿค— (@clementdelangue) 's Twitter Profile Photo

Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scripts,... Full power of open source AI so that everyone all over the world can take advantage of AI progress! Will help debunk some myths Iโ€™m sure too. Thanks

Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scripts,... 

Full power of open source AI so that everyone all over the world can take advantage of AI progress! Will help debunk some myths Iโ€™m sure too. 

Thanks
Inception Labs (@inceptionailabs) 's Twitter Profile Photo

We are excited to introduce Mercury, the first commercial-grade diffusion large language model (dLLM)! dLLMs push the frontier of intelligence and speed with parallel, coarse-to-fine text generation.

Engineering at Meta (@fb_engineering) 's Twitter Profile Photo

Meta and NVIDIA have teamed up to supercharge vector search on GPUs by integrating NVIDIA cuVS into Faiss v1.10, Metaโ€™s open-source library for similarity search. This collaboration brings groundbreaking performance improvements: ๐Ÿ”น IVF indexing: NVIDIA cuVS boosts build times

Meta and NVIDIA have teamed up to supercharge vector search on GPUs by integrating NVIDIA cuVS into Faiss v1.10, Metaโ€™s open-source library for similarity search. This collaboration brings groundbreaking performance improvements:

๐Ÿ”น IVF indexing: NVIDIA cuVS boosts build times
Yann LeCun (@ylecun) 's Twitter Profile Photo

Rob Fergus is the new head of Meta-FAIR! FAIR is refocusing on Advanced Machine Intelligence: what others would call human-level AI or AGI. linkedin.com/posts/rob-fergโ€ฆ

Weixin Liang (@liang_weixin) 's Twitter Profile Photo

๐ŸŽ‰ Excited to share: "๐Œ๐ข๐ฑ๐ญ๐ฎ๐ซ๐ž-๐จ๐Ÿ-๐“๐ซ๐š๐ง๐ฌ๐Ÿ๐จ๐ซ๐ฆ๐ž๐ซ๐ฌ (๐Œ๐จ๐“)" has been officially accepted to TMLR (March 2025) and the code is now open-sourced! ๐Ÿ“Œ GitHub repo: github.com/facebookresearโ€ฆ ๐Ÿ“„ Paper: arxiv.org/abs/2411.04996 How can we reduce pretraining costs for

๐ŸŽ‰ Excited to share: "๐Œ๐ข๐ฑ๐ญ๐ฎ๐ซ๐ž-๐จ๐Ÿ-๐“๐ซ๐š๐ง๐ฌ๐Ÿ๐จ๐ซ๐ฆ๐ž๐ซ๐ฌ (๐Œ๐จ๐“)" has been officially accepted to TMLR (March 2025) and the code is now open-sourced!

๐Ÿ“Œ GitHub repo: github.com/facebookresearโ€ฆ
๐Ÿ“„ Paper: arxiv.org/abs/2411.04996

How can we reduce pretraining costs for
Gargi Ghosh (@gargighosh) 's Twitter Profile Photo

Factuality and hallucination is a big problem in foundation models. We demonstrate that itโ€™s possible to improve long form factuality by 35% without losing helpfulness/ ability to provide detailed response with long CoT, new reward for GRPO and tricks to stop reward hacking

AI at Meta (@aiatmeta) 's Twitter Profile Photo

๐Ÿ† We're thrilled to announce that Meta FAIRโ€™s Brain & AI team won 1st place at the prestigious Algonauts 2025 brain modeling competition. Their 1B parameter model, TRIBE (Trimodal Brain Encoder), is the first deep neural network trained to predict brain responses to stimuli

Gargi Ghosh (@gargighosh) 's Twitter Profile Photo

New research from FAIR- Active Reading: a framework to learn a given set of material with self-generated learning strategies for generalized and expert domains(such as Finance). Absorb significantly more knowledge than vanilla finetuning and usual data augmentations strategies