Vincent Weisser (@vincentweisser) 's Twitter Profile
Vincent Weisser

@vincentweisser

ceo @primeintellect / decentralized agi & science

ID: 427766081

linkhttp://vincentweisser.com calendar_today03-12-2011 23:18:48

9,9K Tweet

19,19K Followers

2,2K Following

moons (@moonsandhues) 's Twitter Profile Photo

⢀⣀⣀⣀⣀⠀⠀⠀⠀⠀⠀⠀⠀⢀⣀⡀⠀⠀ ⠘⢿⣿⡿⠿⢿⣦⡐⠈⡀⢂⡖⠿⠻⢽⣿⣿⠃ ⠀⠈⣿⣿⣮⠂⡝⡻⣆⡰⣻⡣⠜⡧⢾⣞⠁⠀ ⠀⠀⠙⠛⠿⢶⡤⢉⢮⠝⠢⢠⡵⠾⠛⠁⠀⠀ ⠀⠀⠀⣠⣾⣷⡰⠥⣪⡞⡬⡔⡸⣳⡄⠀⠀⠀ ⠀⠀⠀⠈⢿⣿⣾⢷⡋⠙⣶⣷⣟⣯⠃⠀⠀⠀ ⠀⠀⠀⠀⠈⠙⠛⠋⠀⠀⠉⠛⠋⠁

Vincent Weisser (@vincentweisser) 's Twitter Profile Photo

Awesome work by Mario Sieg to accelerate quantization of pseudo-gradients in decentralized training settings like DiLoCo - already integrated in pccl (prime collective communication library)

Casper Hansen (@casper_hansen_) 's Twitter Profile Photo

will brown cooking up some amazing async RL things! dev branch reveals async off-policy steps, which can easily give you 2-3x because of higher GPU utilization👀

<a href="/willccbb/">will brown</a> cooking up some amazing async RL things! dev branch reveals async off-policy steps, which can easily give you 2-3x because of higher GPU utilization👀
will brown (@willccbb) 's Twitter Profile Photo

i'm teaming up with Kyle Corbitt from openpipe to teach a class about agents + RL :) we'll be teaching the class on Maven 🏛 starting june 16. as far as we know, this is the first course of its kind anywhere to bridge RL + LLM agents, and we’re really excited to share some of our

i'm teaming up with <a href="/corbtt/">Kyle Corbitt</a> from openpipe to teach a class about agents + RL :)

we'll be teaching the class on <a href="/MavenHQ/">Maven 🏛</a> starting june 16. as far as we know, this is the first course of its kind anywhere to bridge RL + LLM agents, and we’re really excited to share some of our
will brown (@willccbb) 's Twitter Profile Photo

fun fact: @primeintellect has around 20 employees total everyone is exceptional at what they do. you have a lot of autonomy, and that comes with a lot of responsibility. we're hiring, but not rapidly. we want someone really, really good for this role. sound like fun?

Riva (@rivatez) 's Twitter Profile Photo

co-led a session with the awesome avwtr at Edge Esmeralda yesterday on what we claim to be the ultimate scientific bottleneck- our own miscalibrated conceptions of what science even is an example this is how we've reduced words/concepts to v narrow meanings. people can

co-led a session with the awesome <a href="/alexvawter/">avwtr</a> at <a href="/EdgeEsmeralda/">Edge Esmeralda</a> yesterday on what we claim to be the ultimate scientific bottleneck-  our own miscalibrated conceptions of what science even is

an example this is how we've reduced words/concepts to v narrow meanings. people can
Justus Mattern (@matternjustus) 's Twitter Profile Photo

another easy to generate verifiable task: ask an LLM to generate a JSON that adheres to a very complex (LLM-generated) pydantic model Coming soon to the most diverse RL dataset out there, along with many other tasks that are not math and coding

another easy to generate verifiable task: ask an LLM to generate a JSON that adheres to a very complex (LLM-generated) pydantic model

Coming soon to the most diverse RL dataset out there, along with many other tasks that are not math and coding
mike64_t (@mike64_t) 's Twitter Profile Photo

A 256x256 matrix memory at 2048 tokens of sequence length can achieve 98.6% retrieval accuracy. With random embeddings, accuracy would saturate at 67% and with optimized embeddigs at 80%. Combined with a 4-layer LSTM decoder and a 1-layer store gate, the accuracy increases to

A 256x256 matrix memory at 2048 tokens of sequence length can achieve 98.6% retrieval accuracy.
With random embeddings, accuracy would saturate at 67% and with optimized embeddigs at 80%.
Combined with a 4-layer LSTM decoder and a 1-layer store gate, the accuracy increases to
Grad (@grad62304977) 's Twitter Profile Photo

Seems like no one saw this either, scraping arxiv manually seems to be the way. Pretty cool paper on rl for creative writing on Qwen3 32B base, and most interestingly it's one author from the Star Writing Team (haven't heard of them). They seem to have access to the 32B base tho

Seems like no one saw this either, scraping arxiv manually seems to be the way. Pretty cool paper on rl for creative writing on Qwen3 32B base, and most interestingly it's one author from the Star Writing Team (haven't heard of them). They seem to have access to the 32B base tho
Johannes Hagemann (@johannes_hage) 's Twitter Profile Photo

ai tpot is missing lots of banger RL arxiv releases every week but Grad isn't. we gotta post the weekly docs he creates for our paper reading group somewhere

ai tpot is missing lots of banger RL arxiv releases every week but <a href="/Grad62304977/">Grad</a> isn't.

we gotta post the weekly docs he creates for our paper reading group somewhere
samsja (@samsja19) 's Twitter Profile Photo

async RL is faster that synchronous counterpart. this might be the first time in ML history where an algorithm is naturally async at scale. we realized two things at prime 6 months ago: * RL will be as much compute intensive as pretraining, pushing frontier capability * for

async RL is faster that synchronous counterpart.
 
this might be the first time in ML history where an algorithm is naturally async at scale.

we realized two things at prime 6 months ago:

* RL will be as much compute intensive as pretraining, pushing frontier capability

* for