Michel Olvera (@michelolzam) 's Twitter Profile
Michel Olvera

@michelolzam

PhD in Computer Science @Inria_Nancy. Researcher in audio at @telecomparis.

ID: 1054248180353785858

linkhttp://molveraz.com calendar_today22-10-2018 05:48:47

386 Tweet

207 Followers

1,1K Following

Vivek Verma (@vcubingx) 's Twitter Profile Photo

IT'S FINALLY OUT! After months of work, I'm happy to announce that the first video in a series on visualizing deep learning is out. It's on visualizing the structure of a neural network. RTs appreciated :) youtu.be/UOvPeC8WOt8

IT'S FINALLY OUT!

After months of work, I'm happy to announce that the first video in a series on visualizing deep learning is out. It's on visualizing the structure of a neural network.

RTs appreciated :)

youtu.be/UOvPeC8WOt8
FaRo (@faroit) 's Twitter Profile Photo

Happy to see that DeMask built with asteroid won the first place of the PyTorch summer #hackathon! The model allows you to enhance muffled speech when wearing facemask Demo: youtu.be/QLf10Uqu8Yk 👋 to our team: Manuel Pariente Michel Olvera @_jonashaag and Samuel Cornell

Manuel Pariente (@mnlpariente) 's Twitter Profile Photo

Woohoo 🤩 We won the @Pytorch #hackathon with DeMask - built with Asteroid ! Truly surprised and humbled, but also very proud of our team (Samuele Cornell, FaRo, Michel Olvera and @_jonashaag). Project page: devpost.com/software/aster… Demo: youtube.com/watch?v=QLf10U… Thanks !

Eduardo Fonseca (@edfonseca_) 's Twitter Profile Photo

🔊Happy to announce FSD50K: the new open dataset of human-labeled sound events! Over 51k Freesound audio clips, totalling over 100h of audio manually labeled using 200 classes drawn from the AudioSet Ontology. Paper: arxiv.org/pdf/2010.00475… Dataset: doi.org/10.5281/zenodo…

🔊Happy to announce FSD50K: the new open dataset of human-labeled sound events! Over 51k Freesound audio clips, totalling over 100h of audio manually labeled using 200 classes drawn from the AudioSet Ontology.

Paper: arxiv.org/pdf/2010.00475…
Dataset: doi.org/10.5281/zenodo…
National Geographic (@natgeo) 's Twitter Profile Photo

Día de los Muertos, or #DayoftheDead, unfolds in an explosion of color and joy demonstrating love and respect for lost loved ones on.natgeo.com/3kPrIh9

@danstowell@mastodon.social (@mclduk) 's Twitter Profile Photo

New preprint: "Computational bioacoustics with deep learning: a review and roadmap" arxiv.org/abs/2112.06725 #bioacoustics #machinelearning

Daniela Robles @daniela-oaks.bsky.social (@daniela_oaks) 's Twitter Profile Photo

Hi guys / Hola amigos / Bom dia, Are you a researcher in a low-and-middle income country and working in #melanoma (any aspect of it)? Please get in touch, I would like to know more about what you're doing and your current challenges 😊 Please RT!!!

arXiv Sound (@arxivsound) 's Twitter Profile Photo

``ReCLAP: Improving Zero Shot Audio Classification by Describing Sounds,'' Sreyan Ghosh, Sonal Kumar, Chandra Kiran Reddy Evuru, Oriol Nieto, Ramani Duraiswami, Dinesh Manocha, ift.tt/OgxJruG

Kristina Ulicna, PhD 👩‍💻 (@kristinaulicna) 's Twitter Profile Photo

Important reminder 🚨: with the surge of "all 17 papers from my lab got accepted at NeurIPS 🤩" posts today, just remember that it's perfectly fine if your paper got rejected 🙅‍♀️ / if you haven't even submitted 🤷‍♀️. You can do awesome research even without flashy publications! 📝

Hugo (@mldhug) 's Twitter Profile Photo

You want to give audio abilities to your VLM without compromising its vision performance? You want to align your audio encoder with a pretrained image encoder without suffering from the modality gap? Check our #NeurIPS2024 paper with Michel Olvera Stéphane LATHUILIÈRE and Slim Essid

You want to give audio abilities to your VLM without compromising its vision performance?  You want to align your audio encoder with a pretrained image encoder without suffering from the modality gap? 
Check our #NeurIPS2024 paper with <a href="/michelolzam/">Michel Olvera</a> <a href="/Steph_lat/">Stéphane LATHUILIÈRE</a>  and Slim Essid
Michel Olvera (@michelolzam) 's Twitter Profile Photo

Great talk today by Haohe Liu at the ADASP group on Latent Diffusion Models (LDMs) as versatile audio decoder! Walked us through diffusion basics, AudioLDM for text-to-audio, audio quality enhancement, and neural codecs!

Great talk today by <a href="/LiuHaohe/">Haohe Liu</a> at the <a href="/tp_adasp/">ADASP</a> group on Latent Diffusion Models (LDMs) as versatile audio decoder! Walked us through diffusion basics, AudioLDM for text-to-audio, audio quality enhancement, and neural codecs!
DCASE Workshop (@dcase_workshop) 's Twitter Profile Photo

Thank you for participating in #DCASE2024 workshop. Your participation made it a meaningful experience, and we look forward to staying connected.

Thank you for participating in #DCASE2024 workshop. Your participation made it a meaningful experience, and we look forward to staying connected.
Magdalena Fuentes (@mfu3ntes) 's Twitter Profile Photo

🔊🎶 MIR folks, mirdata 0.3.9 is out‼️ ISMIR Conference #ISMIR2024 6 new loaders, for a total of 58 dataset loaders! This time, we made mirdata MUCH lighter! Install Time: 50.6s ⏩slashed by 31% to 34.6s Size: 95MB ⏬ shrunk by 98% to 1.9MB!! Datasets: mirdata.readthedocs.io/en/stable/sour…

🔊🎶 MIR folks, mirdata 0.3.9 is out‼️ <a href="/ISMIRConf/">ISMIR Conference</a> #ISMIR2024 

6 new loaders, for a total of 58 dataset loaders! 

This time, we made mirdata MUCH lighter!
Install Time: 50.6s ⏩slashed by 31% to 34.6s
Size: 95MB ⏬ shrunk by 98% to 1.9MB!!

Datasets: mirdata.readthedocs.io/en/stable/sour…
Wieland Brendel (@wielandbr) 's Twitter Profile Photo

🎉 New Pre-print! 🎉 Do CLIP models truly generalize to new, out-of-domain (OOD) images, or are they only doing well because they’ve been exposed to these domains in training? Our latest study reveals that CLIP’s ability to “generalize OOD” may be more limited than previously

🎉 New Pre-print! 🎉
Do CLIP models truly generalize to new, out-of-domain (OOD) images, or are they only doing well because they’ve been exposed to these domains in training? Our latest study reveals that CLIP’s ability to “generalize OOD” may be more limited than previously
Hugo (@mldhug) 's Twitter Profile Photo

If you want to learn more about audio-visual alignment and how to use it to give audio abilities to your VLM, stop by our NeurIPS Conference poster #3602 (East exhibit hall A-C) tomorrow at 11am!