McAuley Lab UCSD (@mcauleylabucsd) 's Twitter Profile
McAuley Lab UCSD

@mcauleylabucsd

We're the McAuley lab @ucsd_cse with PI Prof. Julian McAuley!

We work and tweet about cool #MachineLearning and #NLProc applications 🧠🤖

ID: 1458563571810390021

linkhttps://cseweb.ucsd.edu/~jmcauley/ calendar_today10-11-2021 22:34:03

16 Tweet

273 Followers

26 Following

Hao-Wen (Herman) Dong 董皓文 (@hermanhwdong) 's Twitter Profile Photo

Happy to share that our paper "Deep Performer: Score-to-Audio Music Performance Synthesis" has been accepted to IEEE ICASSP 2022! 🥳 This joint work with Cong Zhou, Taylor Berg-Kirkpatrick and Julian McAuley (McAuley Lab UCSD) is based on my internship work at Dolby last summer. 🎶

Noveen Sachdeva (@noveens97) 's Twitter Profile Photo

Conventional #recsys wisdom: "better to go wide than deep". Our paper: go infinitely-wide, compute the solution in closed-form with a single hyper-parameter, and considerably beat all SoTA. Furthermore, can you get the same performance with just 500 fake users? Yes! A thread 🧵

Conventional #recsys wisdom: "better to go wide than deep". Our paper: go infinitely-wide, compute the solution in closed-form with a single hyper-parameter, and considerably beat all SoTA. Furthermore, can you get the same performance with just 500 fake users? Yes! A thread 🧵
Bodhisattwa Majumder (@mbodhisattwa) 's Twitter Profile Photo

Announcing with shaky hands and much delight: Our "conversational critiquing" paper is selected for "Highlights of ACM RecSys '22". 🎉 Didn't know before what it is like to be among the bests of a conf ~ ACM RecSys Paper: bit.ly/3TMX891 Shuyang Li McAuley Lab UCSD 🤩

Announcing with shaky hands and much delight:
Our "conversational critiquing" paper is selected for "Highlights of ACM RecSys '22". 🎉 Didn't know before what it is like to be among the bests of a conf ~ <a href="/ACMRecSys/">ACM RecSys</a>
Paper: bit.ly/3TMX891 

<a href="/ShuyangLi2/">Shuyang Li</a> <a href="/McAuleyLabUCSD/">McAuley Lab UCSD</a> 🤩
UCSD Engineering (@ucsdjacobs) 's Twitter Profile Photo

Researchers UC San Diego developed algorithms to rid speech generated by online bots of offensive language, on social media and elsewhere. For more stories about the #UCEngineer impact on #CyberSecurity, visit ucal.us/engineersweek #Eweek2023 #UCEngineer

Researchers <a href="/UCSanDiego/">UC San Diego</a> developed algorithms to rid speech generated by online bots of offensive language, on social media and elsewhere.
For more stories about the #UCEngineer impact on #CyberSecurity, visit ucal.us/engineersweek
#Eweek2023 #UCEngineer
Noveen Sachdeva (@noveens97) 's Twitter Profile Photo

Ecstatic to join @DeepMind as a research intern for the summer -- looking forward to new friends and being surrounded by the smartest of smartest 🦾 Please DM me if you're around MTV-CE, let's go for a coffee ☕️

Noveen Sachdeva (@noveens97) 's Twitter Profile Photo

Highly grateful! Definitely recommend the streamlined publication experience Transactions on Machine Learning Research For people intersted in data distillation, do checkout our survey - it designed to be to-the-point, and does not require a lot of prerequisite knowledge. Any feedback is highly appreciated!

Highly grateful! Definitely recommend the streamlined publication experience <a href="/TmlrOrg/">Transactions on Machine Learning Research</a> 

For people intersted in data distillation, do checkout our survey - it designed to be to-the-point, and does not require a lot of prerequisite knowledge. 

Any feedback is highly appreciated!
Zhankui He (@zhankuihe) 's Twitter Profile Photo

🤖️ Are LLMs good Conversational Recommender Systems (CRS) ? We (McAuley Lab UCSD and Netflix Research) let LLMs generate movie names directly in response to natural-language user requests. Key observations in the experiments:

🤖️ Are LLMs good Conversational Recommender Systems (CRS) ? 
We (<a href="/McAuleyLabUCSD/">McAuley Lab UCSD</a> and <a href="/NetflixResearch/">Netflix Research</a>) let LLMs generate movie names directly in response to natural-language user requests. Key observations in the experiments:
AK (@_akhaliq) 's Twitter Profile Photo

Farzi Data: Autoregressive Data Distillation paper page: huggingface.co/papers/2310.09… study data distillation for auto-regressive machine learning tasks, where the input and output have a strict left-to-right causal structure. More specifically, we propose Farzi, which summarizes an

Farzi Data: Autoregressive Data Distillation

paper page: huggingface.co/papers/2310.09…

study data distillation for auto-regressive machine learning tasks, where the input and output have a strict left-to-right causal structure. More specifically, we propose Farzi, which summarizes an
Zachary Novack @ICLR2025 🇸🇬 (@zacknovack) 's Twitter Profile Photo

Lead sheets concisely describe music, but can we improve their compressive ability w.r.t. the original score? Check out our new work - Unsupervised Lead Sheet Generation via Semantic Compression 📖 arxiv.org/abs/2310.10772 w/n i k i t a Taylor Berg-Kirkpatrick McAuley Lab UCSD 1/n

Yupeng Hou (@yupenghou97) 's Twitter Profile Photo

Our paper on interesting findings about “LLMs & RecSys” has just been accepted as a full paper in #ecir2024 The most delightful thing is that we got really high-quality, detailed, and constructive reviews. Thanks reviewers from ECIR2024 ! A thread 🧵 github.com/RUCAIBox/LLMRa…

Zachary Novack @ICLR2025 🇸🇬 (@zacknovack) 's Twitter Profile Photo

Fine-grained control/editing in text-to-music diffusion models w/NO TRAINING? Presenting DITTO: Diffusion Inference-Time T-Optimization for Music Generation 📖:arxiv.org/abs/2401.12179 🎹:ditto-music.github.io/web/ w/McAuley Lab UCSD Taylor Berg-Kirkpatrick Nicholas J. Bryan🧵

Jeff Dean (@jeffdean) 's Twitter Profile Photo

Gemini 1.5 Pro - A highly capable multimodal model with a 10M token context length Today we are releasing the first demonstrations of the capabilities of the Gemini 1.5 series, with the Gemini 1.5 Pro model. One of the key differentiators of this model is its incredibly long

Gemini 1.5 Pro - A highly capable multimodal model with a 10M token context length

Today we are releasing the first demonstrations of the capabilities of the Gemini 1.5 series, with the Gemini 1.5 Pro model.  One of the key differentiators of this model is its incredibly long
Noveen Sachdeva (@noveens97) 's Twitter Profile Photo

Q: Can we pre-train LLMs efficiently (and better?) via data pruning? A: Yes! Q: How? A: (secret) Prompt LLMs for data quality 🤫 Check out our latest work Google DeepMind - “How to Train Data-Efficient LLMs” 📖 arxiv.org/abs/2402.09668 An expensive thread 🧵(RTs appreciated!)

Q: Can we pre-train LLMs efficiently (and better?) via data pruning?
A: Yes!

Q: How?
A: (secret) Prompt LLMs for data quality 🤫

Check out our latest work <a href="/GoogleDeepMind/">Google DeepMind</a> - “How to Train Data-Efficient LLMs”
📖 arxiv.org/abs/2402.09668

An expensive thread 🧵(RTs appreciated!)
Yupeng Hou (@yupenghou97) 's Twitter Profile Photo

🚀 Releasing Amazon Reviews 2023 dataset! With *500+M* user reviews, *48+M* items, *60+B* tokens, all from 33 categories, Amazon Reviews, one of the largest, most widely-used review dataset has come to its fourth generation. A thread 🧵 amazon-reviews-2023.github.io