Daniela Gottesman (@dhgottesman) 's Twitter Profile
Daniela Gottesman

@dhgottesman

ID: 1473492874352234497

calendar_today22-12-2021 03:17:46

13 Tweet

27 Followers

18 Following

Mor Geva (@megamor2) 's Twitter Profile Photo

Do you have a "tell" when you are about to lie? We find that LLMs have “tells” in their internal representations which allow estimating how knowledgeable a model is about an entity 𝘣𝘦𝘧𝘰𝘳𝘦 it generates even a single token. Paper: arxiv.org/abs/2406.12673… 🧵 Daniela Gottesman

Do you have a "tell" when you are about to lie?

We find that LLMs have “tells” in their internal representations which allow estimating how knowledgeable a model is about an entity 𝘣𝘦𝘧𝘰𝘳𝘦 it generates even a single token.

Paper: arxiv.org/abs/2406.12673… 🧵

<a href="/dhgottesman/">Daniela Gottesman</a>
Tomer Porian (@tomerporian) 's Twitter Profile Photo

🧵1/8 We resolve the discrepancy between the compute optimal scaling laws of Kaplan (exponent 0.88, Figure 14, left) et al. and Hoffmann et al. (“Chinchilla”, exponent 0.5). Paper: arxiv.org/abs/2406.19146 Data + Code: github.com/formll/resolvi…

🧵1/8 We resolve the discrepancy between the compute optimal scaling laws of Kaplan (exponent 0.88, Figure 14, left) et al. and Hoffmann et al. (“Chinchilla”, exponent 0.5).
Paper: arxiv.org/abs/2406.19146
Data + Code: github.com/formll/resolvi…
Mor Geva (@megamor2) 's Twitter Profile Photo

In Hebrew, we have an idiom "one in the mouth, one in the heart" which means that there is a gap between what someone says versus what they think. Daniela Gottesman's recent work showed with a simple probe (KEEN) that this behavior often happens in LLMs -->

In Hebrew, we have an idiom "one in the mouth, one in the heart" which means that there is a gap between what someone says versus what they think.

<a href="/dhgottesman/">Daniela Gottesman</a>'s recent work showed with a simple probe (KEEN) that this behavior often happens in LLMs --&gt;