Dipam Chakraborty (@__dipam__) 's Twitter Profile
Dipam Chakraborty

@__dipam__

Machine Learning Engineer at @h2oai 🤖

Tweets about topics that interest me in Deep Learning/Reinforcement Learning

ID: 1199391111761915904

linkhttp://dipam.in calendar_today26-11-2019 18:15:06

1,1K Tweet

196 Followers

690 Following

Philipp Singer (@ph_singer) 's Twitter Profile Photo

Whenever you are contemplating participating in Kaggle competitions and you might have heard someone say it is too far-fetched from practical data science work, consider this example: In the recent Science LLM competition participants learned among many other things: - How to

Dipam Chakraborty (@__dipam__) 's Twitter Profile Photo

Yep, gradient checkpointing is probably the most underrated memory saving trick that just works. Definitely should be known well by all GPU poor folks.

Sergey Levine (@svlevine) 's Twitter Profile Photo

Brand new CS285: Deep RL lecture on RL for sequence models and language models: youtu.be/egJgDbe5oaM?fe… Figured it was about time to add this lecture to the course🙂

Dipam Chakraborty (@__dipam__) 's Twitter Profile Photo

While I love LlamaIndex, but I'm bearish on architecture based RAG in the long term, this feels like brittle pre-ML over engineering that will soon be replaced by one architecture to rule them all.

Linas Beliūnas (@linasbeliunas) 's Twitter Profile Photo

Sergey Brin is worth $105 billion yet he was a core contributor on the Gemini AI technical paper, coding basically every day. Legend.

Sergey Brin is worth $105 billion yet he was a core contributor on the Gemini AI technical paper, coding basically every day.

Legend.
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

Idea: safeLinux. All the same programs you know and love but now upgraded with safety to stop bad actors right in their tracks. $ ls I'm sorry, I cannot list the files in this directory because one or more files may contain unsafe content. Can I help you with anything else? 😅

John Carmack (@id_aa_carmack) 's Twitter Profile Photo

Every solar panel is like a giant pixel in a camera sensor, turning photons into a voltage. With billions of them across the world, it is interesting to imagine what they could collectively “see” of the sun and atmosphere, even though they have no focusing machinery. Conversely,

Dipam Chakraborty (@__dipam__) 's Twitter Profile Photo

I'm a 100% EV proponent, but a friend pointed out an interesting paradox about this price dynamic. Batteries get cheaper so the resale value of EV gets lower, whereas ICE cars are getting more expensive so they are also holding resale value. Weird market dynamics.

Yi Tay (@yitayml) 's Twitter Profile Photo

not true, especially for language. if you trained a large & deep MLP language model with no self-attention, no matter how much data you'll feed it you'll still be lacking behind a transformer (with much less data). will it get to the same point? i don't think so. your tokens

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

LLM model size competition is intensifying… backwards! My bet is that we'll see models that "think" very well and reliably that are very very small. There is most likely a setting even of GPT-2 parameters for which most people will consider GPT-2 "smart". The reason current