Rick Lamers (@ricklamers) 's Twitter Profile
Rick Lamers

@ricklamers

πŸ‘¨β€πŸ’» AI Research & Engineering @GroqInc. Occasional angel investor. I publish technical resources about LLMs every week. Opinions are my own.

ID: 57274933

linkhttps://codingwithintelligence.com/ calendar_today16-07-2009 07:48:12

2,2K Tweet

5,5K Followers

668 Following

Rick Lamers (@ricklamers) 's Twitter Profile Photo

This incredible work reveals mitigating a hidden bias could contribute to improving long context performance, potentially opening up the path to a formulation that can direct attention equally to tokens deeper in the sequence. πŸ‘ Francesco Pappone

Rick Lamers (@ricklamers) 's Twitter Profile Photo

Come chat to Aarush Sah about OpenBench! πŸ”₯ Ask him about: eval saturation, harness influence, LLM-as-a-judge, up-and-coming evals, which evals are useful, open source vs proprietary models on various evals, and much much more!

Rick Lamers (@ricklamers) 's Twitter Profile Photo

Core takeaway from open-sci-ref 0.01: data quality > scale. A 1.7B model trained on Nemotron-CC-HQ for 1T tokens matches SmolLM2-1.7B trained on ~11T. Further confirms the value of curating high quality (open) datasets.

Rick Lamers (@ricklamers) 's Twitter Profile Photo

RL finds a way, intelligence is finding the path of least resistance. We just need to make sure the path of least resistance is a useful one.