Raj Palleti (@rajpalleti314) 's Twitter Profile
Raj Palleti

@rajpalleti314

ID: 1693689759900786688

calendar_today21-08-2023 18:21:31

224 Tweet

127 Followers

71 Following

Shantanu Sharma (@shantanu) 's Twitter Profile Photo

Good read: The Leaderboard Illusion: alphaxiv.org/abs/2504.20879 Big Tech commercially dependent on marketing model performance for revenues putting their best models out on Chatbot Arena is not surprising. I would argue against prohibiting score retraction after submission and

alphaXiv (@askalphaxiv) 's Twitter Profile Photo

Are LLM leaderboards no longer trustworthy? cohere's deep dive reveals how Chatbot Arena scores can be gamed 🔍 Providers test 10–27 private models & submit the best 📊 Proprietary models receive 2–3× more data ⚠️ Rankings may reflect overfitting Trending on alphaXiv 📈

Are LLM leaderboards no longer trustworthy?

<a href="/cohere/">cohere</a>'s deep dive reveals how Chatbot Arena scores can be gamed

🔍 Providers test 10–27 private models &amp; submit the best
📊 Proprietary models receive 2–3× more  data
⚠️ Rankings may reflect overfitting

Trending on alphaXiv 📈
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

🚨Bright week for agents and representation learning, notably including X-Fusion’s remarkable progress in multimodal capabilities 🚀 Check out the top 10 papers for the week👇 - From LLM Reasoning to Autonomous AI Agents: A Comprehensive Review - Reinforcement Learning for

🚨Bright week for agents and representation learning, notably including X-Fusion’s remarkable progress in multimodal capabilities 🚀

Check out the top 10 papers for the week👇

- From LLM Reasoning to Autonomous AI Agents: A Comprehensive Review
- Reinforcement Learning for
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

The most comprehensive survey (800+ papers) on LLM safety is here 👀 It highlights how agents, with tools, memory & env access, dramatically expand the attack surface This paper offers a "Full-Stack" analysis of these critical, often overlooked, risks Trending on alphaXiv 🚀

The most comprehensive survey (800+ papers) on LLM safety is here 👀

It highlights how agents, with tools, memory &amp; env access, dramatically expand the attack surface

This paper offers a "Full-Stack" analysis of these critical, often overlooked, risks

Trending on alphaXiv 🚀
John Bohannon (@bohannon_bot) 's Twitter Profile Photo

Trying out alphaXiv today. Loving how you can ask scientists questions within the context of their papers. The authors of the Absolute Zero paper (self-learning RL) @AndrewZ45732491 Shenzhi Wang🌟 and Qingyun Wu are kindly answering questions at alphaxiv.org/abs/2505.03335…

alphaXiv (@askalphaxiv) 's Twitter Profile Photo

🚨Dive into Absolute Zero and RM-R1's major breakthrough in reasoning, unveiling another exciting week for AI.🚀 Check out the top 10 papers for the week👇 - Absolute Zero: Reinforced Self-play Reasoning with Zero Data - RM-R1: Reward Modeling as Reasoning - ZeroSearch:

🚨Dive into Absolute Zero and RM-R1's major breakthrough in reasoning, unveiling another exciting week for AI.🚀

Check out the top 10 papers for the week👇

- Absolute Zero: Reinforced Self-play Reasoning with Zero Data
- RM-R1: Reward Modeling as Reasoning
- ZeroSearch:
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

Can reward models reason? RM-R1 frames reward modeling as a reasoning task—with structured rubrics, long-form evaluations, and verifiable justifications 🧠 +13.8% accuracy on benchmarks & increased interpretability 📊 32B RM-R1 beats GPT-4o & Claude Trending on alphaXiv 📈

Can reward models reason?

RM-R1 frames reward modeling as a reasoning task—with structured rubrics, long-form evaluations, and verifiable justifications

🧠 +13.8% accuracy on benchmarks &amp; increased interpretability
📊 32B RM-R1 beats GPT-4o &amp; Claude

Trending on alphaXiv 📈
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

"The S in MCP stands for Security" See this new review of MCP and its security considerations Key takeaways: ➡️ Security threats exist in ALL phases: creation, operation & update ➡️ Highlights risks of unofficial auto-installers & community servers

"The S in MCP stands for Security"

See this new review of MCP and its security considerations

Key takeaways: 
➡️ Security threats exist in ALL phases: creation, operation &amp; update 
➡️ Highlights risks of unofficial auto-installers &amp; community servers
Percy Liang (@percyliang) 's Twitter Profile Photo

What would truly open-source AI look like? Not just open weights, open code/data, but *open development*, where the entire research and development process is public *and* anyone can contribute. We built Marin, an open lab, to fulfill this vision:

What would truly open-source AI look like? Not just open weights, open code/data, but *open development*, where the entire research and development process is public *and* anyone can contribute. We built Marin, an open lab, to fulfill this vision:
Zhongpai Gao (@zhongpaigao) 's Twitter Profile Photo

I just created Gaussian-Splatting community on alphaXiv for sharing the latest advances in this area. Please join the community for discussions 👏

I just created Gaussian-Splatting community on <a href="/askalphaxiv/">alphaXiv</a> for sharing the latest advances in this area. Please join the community for discussions 👏
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

Introducing Claude 4 Sonnet for understanding arXiv papers 🚀 Highlight any section of a paper to ask questions and “@” other papers to quickly add to context and compare results, benchmarks, etc.

alphaXiv (@askalphaxiv) 's Twitter Profile Photo

🚨Prompt-to-A* Publication has been achieved 🤖🔬A fully AI-generated research paper has been accepted to the main conference of ACL 2025! Intology's research agent, Zochi, discovered and implemented a state-of-the-art jailbreaking attack on LLMs, which has been accepted to

alphaXiv (@askalphaxiv) 's Twitter Profile Photo

This is pretty remarkable – AI systems learning to self-improve We're seeing a wave of research where AI isn't just learning from human feedback, it's starting to figure out how to improve itself using its own internal signals A subtle but profound shift.

This is pretty remarkable – AI systems learning to self-improve

We're seeing a wave of research where AI isn't just learning from human feedback, it's starting to figure out how to improve itself using its own internal signals

A subtle but profound shift.
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

"Can Large Reasoning Models Self-Train?" A brilliant paper from CMU showing LLMs can improve at math reasoning WITHOUT human labels - just learning from their own consistency. Early results rival models trained on ground-truth answers.

"Can Large Reasoning Models Self-Train?"

A brilliant paper from CMU showing LLMs can improve at math reasoning WITHOUT human labels - just learning from their own consistency.

Early results rival models trained on ground-truth answers.
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

🚨There’s a new ceiling for efficient reasoning with the rise of Learning to Reason without External Rewards, along with AgriFM pushing the boundaries of AI to even agriculture🚀 Check out the top 10 papers for the week👇 - Paper2Poster: Towards Multimodal Poster Automation

🚨There’s a new ceiling for efficient reasoning with the rise of Learning to Reason without External Rewards, along with AgriFM pushing the boundaries of AI to even agriculture🚀

Check out the top 10 papers for the week👇

- Paper2Poster: Towards Multimodal Poster Automation
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

Turns out you can jailbreak VLMs with memes A team of Korean researchers showed that pairing harmful prompts with everyday memes makes VLMs way more likely to generate dangerous content than text-only attacks across 50K+ examples

Turns out you can jailbreak VLMs with memes

A team of Korean researchers showed that pairing harmful prompts with everyday memes makes VLMs way more likely to generate dangerous content than text-only attacks across 50K+ examples
Intology (@intologyai) 's Twitter Profile Photo

Introducing the Automated Research community on alphaXiv, created by Intology This Friday, we are hosting Samuel Schmidgall from DeepMind to discuss his work on automated research agents. Come chat with our team and the community! lu.ma/j9r9vf3m 🧵👇