Elad Segal (@eladsegal) 's Twitter Profile
Elad Segal

@eladsegal

Deep Learning Research Engineer @NVIDIA

ID: 885545977

calendar_today16-10-2012 23:23:57

81 Tweet

150 Followers

413 Following

Itay Itzhak (@itay_itzhak_) 's Twitter Profile Photo

📢 New paper alert! 📢 Thrilled to announce `Instructed to Bias: Instruction-Tuned Language Models Exhibit Emergent Cognitive Bias'. Do instruction tuning and RLHF amplify biases in LMs? 🧵 Check it out arxiv.org/abs/2308.00225 W Yonatan Belinkov Gabriel Stanovsky and N. Rosenfeld.

📢 New paper alert! 📢

Thrilled to announce `Instructed to Bias: Instruction-Tuned Language Models Exhibit Emergent Cognitive Bias'.

Do instruction tuning and RLHF amplify biases in LMs? 🧵
Check it out  arxiv.org/abs/2308.00225

W <a href="/boknilev/">Yonatan Belinkov</a> <a href="/GabiStanovsky/">Gabriel Stanovsky</a> and N. Rosenfeld.
Ori Yoran (@oriyoran) 's Twitter Profile Photo

Retrieval-augmented LMs are not robust to irrelevant context. Retrieving entirely irrelevant context can throw off the model, even when the answer is encoded in its parameters! In our new work, we make RALMs more robust to irrelevant context. arxiv.org/abs/2310.01558 🧵[1/7]

Retrieval-augmented LMs are not robust to irrelevant context. Retrieving entirely irrelevant context can throw off the model, even when the answer is encoded in its parameters!

In our new work, we make RALMs more robust to irrelevant context.

arxiv.org/abs/2310.01558

🧵[1/7]
(((ل()(ل() 'yoav))))👾 (@yoavgo) 's Twitter Profile Photo

Hello colleagues and fellows. Over the past few days I was shocked to learn that people in our community don't share what I consider to be basic human values. Please help me restore faith in our community by signing this. forms.gle/2yi1WP9RNSHPHn…

Hello colleagues and fellows. Over the past few days I was shocked to learn that people in our community don't share what I consider to be basic human values. Please help me restore faith in our community by signing this. 

forms.gle/2yi1WP9RNSHPHn…
Visegrád 24 (@visegrad24) 's Twitter Profile Photo

Eyal Waldman is an Israeli billionaire, high-tech magnate (founder of Mellanox) He built R&D centres in the West Bank & Gaza Strip to employ Palestinian developers in order to build better Israeli-Palestinian relations. Hamas murdered his daughter Daniel at the music festival

Eyal Waldman is an Israeli billionaire, high-tech magnate (founder of Mellanox)

He built R&amp;D centres in the West Bank &amp; Gaza Strip to employ Palestinian developers in order to build better Israeli-Palestinian relations. 

Hamas murdered his daughter Daniel at the music festival
Elad Simchayoff (@elad_si) 's Twitter Profile Photo

Watch and Share with the world. A special project by החדשות - N12. This video contains footage taken by the young partygoers at the Nova Music Festival prior to the 7.10 terror attack. You’ll only see a handful of the 260 victims and dozens of those abducted or still missing doing

Ben Bogin (@ben_bogin) 's Twitter Profile Photo

Can we leverage pre-existing coding abilities of LLMs to improve semantic parsing and compositional generalization? 🚨 Our new paper shows dramatic improvements when LLMs are prompted with Python rather than DSLs, along with helpful domain descriptions! bit.ly/code-semparse

Can we leverage pre-existing coding abilities of LLMs to improve semantic parsing and compositional generalization?

🚨 Our new paper shows dramatic improvements when LLMs are prompted with Python rather than DSLs, along with helpful domain descriptions!

bit.ly/code-semparse
Guy Dar (@guy_dar1) 's Twitter Profile Photo

🇲🇽 Excited to share our work was accepted to #NAACL2024 main conference!! 🇲🇽 ICL has been hypothesized to perform GD implicitly in its parameters. But is there good evidence for that? 🧐 Depends what you mean exactly!!

Maor Ivgi (@maorivg) 's Twitter Profile Photo

1/7 🚨 What do LLMs do when they are uncertain? We found that the stronger the LLM, the more it hallucinates and the less it loops! This pattern extends to sampling methods and instruction tuning. 🧵👇 Mor Geva Jonathan Berant Ori Yoran

1/7 🚨 What do LLMs do when they are uncertain? We found that the stronger the LLM, the more it hallucinates and the less it loops! This pattern extends to sampling methods and instruction tuning. 🧵👇
<a href="/megamor2/">Mor Geva</a> <a href="/JonathanBerant/">Jonathan Berant</a> <a href="/OriYoran/">Ori Yoran</a>
Ori Yoran (@oriyoran) 's Twitter Profile Photo

Can AI agents solve realistic, time-consuming web tasks such as “Which gyms near me have fitness classes on the weekend, before 7AM?" We introduce AssistantBench, a benchmark with 214 such tasks. Our new GPT-4 based agent gets just 25% accuracy! assistantbench.github.io

Ben Bogin (@ben_bogin) 's Twitter Profile Photo

📢 New Benchmark: SUPER for Setting UP and Executing tasks from Research repositories Reproducibility is crucial in science. We introduce SUPER to evaluate LLMs' capabilities in autonomously running experiments from research repositories. ⬇️ arxiv.org/pdf/2409.07440

📢 New Benchmark: SUPER for Setting UP and Executing tasks from Research repositories

Reproducibility is crucial in science. We introduce SUPER to evaluate LLMs' capabilities in autonomously running experiments from research repositories. ⬇️

arxiv.org/pdf/2409.07440
Mor Geva (@megamor2) 's Twitter Profile Photo

What's in an attention head? 🤯 We present an efficient framework – MAPS – for inferring the functionality of attention heads in LLMs ✨directly from their parameters✨ A new preprint with Amit Elhelo 🧵 (1/10)

What's in an attention head? 🤯

We present an efficient framework – MAPS – for inferring the functionality of attention heads in LLMs ✨directly from their parameters✨

A new preprint with <a href="/AmitElhelo/">Amit Elhelo</a> 🧵 (1/10)
Mor Geva (@megamor2) 's Twitter Profile Photo

How can we interpret LLM features at scale? 🤔 Current pipelines use activating inputs, which is costly and ignores how features causally affect model outputs! We propose efficient output-centric methods that better predict how steering a feature will affect model outputs. New

Itay Levy (@itayoush) 's Twitter Profile Photo

Very excited about the release of the Llama Nemotron Super 49B model 🚀 #GTC25 Using distillation-based NAS (Puzzle) we achieved 5X throughput gain! After SFT and RL, this model tops reasoning benchmarks among open 70B models

Very excited about the release of the Llama Nemotron Super 49B model 🚀 #GTC25

Using distillation-based NAS (Puzzle) we achieved 5X throughput gain!

After SFT and RL, this model tops reasoning benchmarks among open 70B models
Oleksii Kuchaiev (@kuchaev) 's Twitter Profile Photo

We are excited to release Llama-Nemotron-Ultra! This is a reasoning ON/OFF, dense 253B model. Open weights and post-training data. huggingface.co/nvidia/Llama-3… We started with llama-405B, changed it via NAS pruning then followed by reasoning-focused post-training: SFT + RL in FP8.

We are excited to release Llama-Nemotron-Ultra! This is a reasoning ON/OFF, dense 253B model. Open weights and post-training data. huggingface.co/nvidia/Llama-3… We started with llama-405B, changed it via NAS pruning then followed by reasoning-focused post-training: SFT + RL in FP8.