Marius Miron (@nkundiushuti) 's Twitter Profile
Marius Miron

@nkundiushuti

mediteromanian, signal processing and machine learning, ex-photographer, ex-music blogger, amateur musician, field recordist, cyborg

ID: 426721619

linkhttp://mariusmiron.com calendar_today02-12-2011 16:15:30

1,1K Tweet

522 Followers

517 Following

Masato Hagiwara (@mhagiwara) 's Twitter Profile Photo

🎙️✨ Join me and the researchers behind NatureLM-audio (Marius Miron and David Robinson) as we present a live technical walkthrough on Nov 21 at 5pm GMT (12pm US Eastern / 9am US Pacific)! Interested? Head to our Discord community for more information: discord.com/invite/H2Y532a…

wh (@nrehiew_) 's Twitter Profile Photo

16th highest scored paper at ICLR 2025 with 3(!), 8, 10, 10, 10 tldr: they scale sparse autoencoders to GPT4 and show that interpretability techniques used on toy models can work on larger models too (hmm i wonder who these people who have access to GPT4 activations are!)

16th highest scored paper at ICLR 2025 with 3(!), 8, 10, 10, 10

tldr: they scale sparse autoencoders to GPT4 and show that interpretability techniques used on toy models can work on larger models too 

(hmm i wonder who these people who have access to GPT4 activations are!)
Sander Dieleman (@sedielem) 's Twitter Profile Photo

In arxiv.org/abs/2303.00848, Durk Kingma and Ruiqi Gao had suggested that noise augmentation could be used to make other likelihood-based models optimise perceptually weighted losses, like diffusion models do. So cool to see this working well in practice!

Marco Pasini (@marco_ppasini) 's Twitter Profile Photo

✨ Train language models directly on continuous data - without tokenization ✨ We propose an easy way to train GPT-style autoregressive models on continuous data, without error accumulation. We test it on audio 🔊, but this method can easily work with other modalities 🎆 👇🧵

✨ Train language models directly on continuous data - without tokenization ✨

We propose an easy way to train GPT-style autoregressive models on continuous data, without error accumulation.

We test it on audio 🔊, but this method can easily work with other modalities 🎆

👇🧵
Adriano R. Lameira (@lameira_adriano) 's Twitter Profile Photo

🚨 PhD Alert! 🚨 We’re seeking 2 motivated PhD students to explore the evolutionary origins of language through cutting-edge field research with wild orangutans. 🌍🐒 Please RT 📍 Based @WarwickPsych, w/ fieldwork in Sumatra (Indonesia) 📅 Apply now! 👉findaphd.com/phds/project/w…

swyx (@swyx) 's Twitter Profile Photo

this neurips is really going to be remembered as the "end of pretraining" neurips notes from doctor Noam Brown's talk on scaling test time compute today (thank you Hattie Zhou for organizing)

this neurips is really going to be remembered as the "end of pretraining" neurips

notes from doctor <a href="/polynoamial/">Noam Brown</a>'s talk on scaling test time compute today

(thank you <a href="/oh_that_hat/">Hattie Zhou</a> for organizing)
Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

Paper alert––*Awarded best paper* at NeurIPS workshop on Foundation Model Interventions! 🧵👇 We analyze the (in)abilities of SAEs by relating them to the field of disentangled rep. learning, where limitations of AE based interpretability protocols have been well established!🤯

Ibrahim Alabdulmohsin | إبراهيم العبدالمحسن (@ibomohsin) 's Twitter Profile Photo

🔥Excited to introduce RINS - a technique that boosts model performance by recursively applying early layers during inference without increasing model size or training compute flops! Not only does it significantly improve LMs, but also multimodal systems like SigLIP. (1/N)

🔥Excited to introduce RINS - a technique that boosts model performance by recursively applying early layers during inference without increasing model size or training compute flops! Not only does it significantly improve LMs, but also multimodal systems like SigLIP. 
(1/N)
Tanishq Mathew Abraham, Ph.D. (@iscienceluvr) 's Twitter Profile Photo

Large Language Diffusion Models Introduces LLaDA-8B, a large language diffusion model that pretrained on 2.3 trillion tokens using 0.13 million H800 GPU hours, followed by SFT on 4.5 million pairs. LLaDA 8B surpasses Llama-2 7B on nearly all 15 standard zero/few-shot learning

Large Language Diffusion Models

Introduces LLaDA-8B, a large language diffusion model that pretrained on 2.3 trillion tokens using 0.13 million H800 GPU hours, followed by SFT on 4.5 million pairs. LLaDA 8B surpasses Llama-2 7B  on nearly all 15 standard zero/few-shot learning
Marius Miron (@nkundiushuti) 's Twitter Profile Photo

happy to announce that Biodenoising was accepted at ICASSP 2025. this is essentially the equivalent of speech enhancement for non-human vocalizations. it can be easily used in Python with pip install biodenoising.

Earth Species Project (@earthspecies) 's Twitter Profile Photo

ESP co-founder, Aza Raskin spoke with Kenneth Cukier on the Babbage podcast by The Economist, where he shared how we're leveraging AI to decode animal communication and working toward a future of interspecies understanding 🌍 economist.com/podcasts/2025/…

ESP co-founder, <a href="/aza/">Aza Raskin</a> spoke with <a href="/kncukier/">Kenneth Cukier</a> on the Babbage podcast by <a href="/TheEconomist/">The Economist</a>, where he shared how we're leveraging AI to decode animal communication and working toward a future of interspecies understanding 🌍

economist.com/podcasts/2025/…
Marius Miron (@nkundiushuti) 's Twitter Profile Photo

I am at ICASSP and will be presenting biodenoising on Friday morning.happy to talk to people interested in bioacoustics,cross-domain representation transfer or simply curious about our work at ESP.we released a new version of biodenoising including self-training on your own data

Earth Species Project (@earthspecies) 's Twitter Profile Photo

AVES is now pip-installable 🎉 This self-supervised, transformer-based model is pretrained on large-scale animal vocalization datasets & thanks to our incredible engineering team is now more accessible than ever–ready to run with just a single command. 🔗bit.ly/44uTbhN

AVES is now pip-installable 🎉

This self-supervised, transformer-based model is pretrained on large-scale animal vocalization datasets &amp; thanks to our incredible engineering team is now more accessible than ever–ready to run with just a single command.

🔗bit.ly/44uTbhN
Masato Hagiwara (@mhagiwara) 's Twitter Profile Photo

I'll be at #ICLR2025 in Singapore this Saturday for poster session 6, co-presenting our NatureLM-audio project! If you're around and interested in AI for bioacoustics, ecology, or related areas, would love to meet up — feel free to reach out!

I'll be at #ICLR2025 in Singapore this Saturday for poster session 6, co-presenting our NatureLM-audio project! If you're around and interested in AI for bioacoustics, ecology, or related areas, would love to meet up — feel free to reach out!
Earth Species Project (@earthspecies) 's Twitter Profile Photo

📢 We've open-sourced NatureLM-audio, the first audio-language foundation model for #bioacoustics. Trained on large-scale animal vocalization, human speech & music datasets, the model enables zero-shot classification, detection & querying across diverse species & environments 👇🏽

📢 We've open-sourced NatureLM-audio, the first audio-language foundation model for #bioacoustics.

Trained on large-scale animal vocalization, human speech &amp; music datasets, the model enables zero-shot classification, detection &amp; querying across diverse species &amp; environments 👇🏽
Marius Miron (@nkundiushuti) 's Twitter Profile Photo

daré una charla por videoconferencia mañana a las 2PM hora de Colombia sobre cómo usar inteligencia artificial para decodificar el comportamiento animal Earth Species Project (ESP)

jack morris (@jxmnop) 's Twitter Profile Photo

excited to finally share on arxiv what we've known for a while now: All Embedding Models Learn The Same Thing embeddings from different models are SO similar that we can map between them based on structure alone. without *any* paired data feels like magic, but it's real:🧵