Elman Mansimov (@elmanmansimov) 's Twitter Profile
Elman Mansimov

@elmanmansimov

AI scientist whose work has been acquired by a prestigious American museum

ID: 4710897316

linkhttp://mansimov.io calendar_today05-01-2016 03:07:12

1,1K Tweet

3,3K Followers

719 Following

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

whoa! this works so well based on my initial tests extracting text from academic papers and invoices super fast and cheap as well

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

you's deep research report looks like the most compelling among the alternatives. finally something more engaging than tons of words. looking forward to trying it

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

just tried chatgpt 4.5 it is indeed better at writing and feels more creative than gpt-4o yeah it might be not much better at benchmarks, but feels too early to arrive at a conclusion

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

but ultimately my conclusion after all LLM releases during last few weeks it is becoming exponentially harder to evaluate latest LLM's capabilities human attention span is flat or decreasing, while changes in model capabilities are getting more nuanced and less obvious

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

after seeing a lot of game dev with Claude 3.7 sonnet and cursor tweets I realized that diffusion and generative models of videos is not the right way for AI generated video games better way is to have a great coding LLM model + 3D asset generation + animation

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

Thanks for the deep dive Monk Antony, it inspired me to take a stroll down memory lane! I started developing alignDRAW in May 2015 and submitted the paper on Nov 9th. Here are some early outputs on github from Sep 2015: github.com/mansimov/cap2i… It's cool to see Alec was

Thanks for the deep dive <a href="/monkantony_tez/">Monk Antony</a>, it inspired me to take a stroll down memory lane! 

I started developing alignDRAW in May 2015 and submitted the paper on Nov 9th. Here are some early outputs on github from Sep 2015: github.com/mansimov/cap2i…

It's cool to see Alec was
Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

i keep hearing that with enough search budget (i.e. several different queries, retries, etc.) sparse retrieval (i.e. keyword / bm25 search) can outperform (or at least match) dense retrieval using embeddings in symbolic domains like code and text indeed if you think about it

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

this looks like a big deal esp related to our understanding of what architectural tricks are important yo train neural nets every since batch norm and layer norm were released they became indispensable in the neural nets architectures making it much easier to stabilize training

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

there is something inherently very fun with training single GPU sized models on small datasets on problems and outputs that motivate you once trained, model outputs look very fun and bring special excitement inside you. plus you get to understand tech even better.

FellowshipAI (@fellowshipai) 's Twitter Profile Photo

How It Started and How It’s Going 🫴 alignDRAW (2015) chatGPT 4o (2025) Inspired by tokumei and his post on alignDRAW compared to ChatGPT, here are some fun comparisons between the two models. 📌 Prompt: A toilet seat sits open in the grass field ← alignDRAW (2015) | chatGPT

How It Started and How It’s Going 🫴
alignDRAW (2015) chatGPT 4o (2025)

Inspired by <a href="/tokumei/">tokumei</a> and his post on alignDRAW compared to ChatGPT, here are some fun comparisons between the two models.

📌 Prompt:
A toilet seat sits open in the grass field

← alignDRAW (2015) | chatGPT
Sainbayar Sukhbaatar (@tesatory) 's Twitter Profile Photo

Ten years ago in 2015 we published a paper called End-to-End Memory Networks (arxiv.org/abs/1503.08895). Looking back, this paper had many of the ingredients of current LLMs. Our model was the first language model that completely replaced RNN with attention. It had dot-product

Ten years ago in 2015 we published a paper called End-to-End Memory Networks (arxiv.org/abs/1503.08895). Looking back, this paper had many of the ingredients of current LLMs. Our model was the first language model that completely replaced RNN with attention. It had dot-product
Arab Bank Switzerland (@arabbankch) 's Twitter Profile Photo

🖼Art in the office - Take a virtual tour of our HQ in Geneva and discover the artworks on display! 📸An Airplane Flying Off Into The Distance On A Clear Day by Elman Mansimov

🖼Art in the office - Take a virtual tour of our HQ in Geneva and discover the artworks on display!
📸An Airplane Flying Off Into The Distance On A Clear Day by <a href="/elmanmansimov/">Elman Mansimov</a>
Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

I am attending ICLR in Singapore next week. Would love to meet new people and old friends. DM or email me to organize the meeting.

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

Cursor Agent and Claude has a tendency to over generate code for my tasks Reminding them to succinct and to the point like their life depends on it is a must

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

was locked out of my delta account for silly reasons had to call their representatives to book a flight — took almost 1 and a half hour we take digital world and internet for granted sometimes

Elman Mansimov (@elmanmansimov) 's Twitter Profile Photo

the better observation here is that multiple choice benchmark via ranking is not the right number to publish esp if your model only available via API we should release official numbers as we actually use the model (via generation) rather than ranking with likelihood