Hamed Shirzad (@hamedshirzad13) 's Twitter Profile
Hamed Shirzad

@hamedshirzad13

PhD student @UBC_CS

Interested in Machine Learning on Graphs

ID: 1550221567174488064

linkhttps://www.hamedshirzad.com/ calendar_today21-07-2022 20:50:08

105 Tweet

182 Followers

247 Following

Yi (Joshua) Ren (looking for postdoc @ICLR) (@joshuarenyi) 's Twitter Profile Photo

Poster at Compositional Learning Workshop @ NeurIPS2024 Exploring why simplicity bias in deep learning often aligns with compositionality. A fresh perspective through the lens of learning dynamics. (1/5) 🗓️Sunday, Dec 15 ⏰8:30 - 17:00 📍West Meeting Room 118-120

Poster at Compositional Learning Workshop @ NeurIPS2024
Exploring why simplicity bias in deep learning often aligns with compositionality. A fresh perspective through the lens of learning dynamics. (1/5)

🗓️Sunday, Dec 15
⏰8:30 - 17:00
📍West Meeting Room 118-120
Mark Schmidt (@markschmidtubc) 's Twitter Profile Photo

For those still at NeurIPS, check out my student's posters at the OPT workshop at 3pm: - Line search with constant sharpness. - Proof line search can beat acceleration. - Allowing negative step sizes? - BCD for deep nets? - Normalization in graph neural net codes makes no sense.

Peymàn M. Kiasari (@pkiasari) 's Twitter Profile Photo

If you are interested in generalization in deep learning, you can find our poster in the poster session right now at #AAAI2025 : )

If you are interested in generalization in deep learning, you can find our poster in the poster session right now at #AAAI2025 : )
Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Happy to see sparse attention methods taking off! We've also worked on sparse attention for learning on graphs (Exphormer, Spexphormer papers) — convincing reviewers about sparse attention (and not just flash attention) wasn’t easy.

Kumo (@kumo_ai_team) 's Twitter Profile Photo

What makes Graph Transformers a perfect fit for relational data? In our latest blog, Federico López breaks down how Kumo integrates Graph Transformers to elevate predictive modeling—eliminating the need for complex pipelines and manual feature engineering. kumo.ai/research/graph…

Yi (Joshua) Ren (looking for postdoc @ICLR) (@joshuarenyi) 's Twitter Profile Photo

📢Curious why your LLM behaves strangely after long SFT or DPO? We offer a fresh perspective—consider doing a "force analysis" on your model’s behavior. Check out our #ICLR2025 Oral paper: Learning Dynamics of LLM Finetuning! (0/12)

📢Curious why your LLM behaves strangely after long SFT or DPO?
We offer a fresh perspective—consider doing a "force analysis" on your model’s behavior.
Check out our #ICLR2025  Oral paper:

Learning Dynamics of LLM Finetuning! 

(0/12)
Kazem Meidani (@kazemmeidani) 's Twitter Profile Photo

Can’t attend ICLR 🇸🇬 due to visa issues but Chandan Reddy will have the oral presentation of *LLM-SR*on Friday 👇🏻 + see our new preprint on benchmarking capabilities of LLMs for scientific equation discovery *LLM-SRBench*: arxiv.org/abs/2504.10415

Jure Leskovec (@jure) 's Twitter Profile Photo

🚀 New tutorial on Graph Transformers! We introduce a powerful architecture combining the strengths of GNNs and Transformers for structured data. Learn how GTs open new possibilities for graph learning. kumo.ai/research/intro… #GraphTransformers #MachineLearning #AI #GNN

Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Huge congrats to my labmate Yi (Joshua) Ren and my supervisor Danica Sutherland for receiving an Outstanding Paper Award at ICLR 2026 for their work on Learning Dynamics of LLM Finetuning! So proud to see their amazing research recognized 👏🔥

Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Thank you Graph Signal Processing Workshop 2025 for the opportunity to give the morning keynote! It was a pleasure to present my work on Exphormer and Spexphormer models to such an insightful audience.

Valence Labs (@valence_ai) 's Twitter Profile Photo

1/ At Valence Labs, Recursion's AI research engine, we’re focused on advancing drug discovery outcomes through cutting-edge computational methods Today, we're excited to share our vision for building virtual cells, guided by the predict-explain-discover framework 🧵

1/ At Valence Labs, <a href="/RecursionPharma/">Recursion</a>'s AI research engine, we’re focused on advancing drug discovery outcomes through cutting-edge computational methods

Today, we're excited to share our vision for building virtual cells, guided by the predict-explain-discover framework 🧵
Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Loved working on the TxPert project! It's also exciting to see my PhD work on Graph Transformers (Exphormer model) finding such meaningful application in a critical real-world task.

Valence Labs (@valence_ai) 's Twitter Profile Photo

1/ Introducing TxPert: a new model that predicts transcriptional responses across diverse biological contexts It’s designed to generalize across unseen single-gene perturbations, novel combinations of gene perturbations, and even new cell types 🧵

Ali Behrouz (@behrouz_ali) 's Twitter Profile Photo

What makes attention the critical component for most advances in LLMs and what holds back long-term memory modules (RNNs)? Can we strictly generalize Transformers? Presenting Atlas (A powerful Titan): a new architecture with long-term in-context memory that learns how to

What makes attention the critical component for most advances in LLMs and what holds back long-term memory modules (RNNs)? Can we strictly generalize Transformers?

Presenting Atlas (A powerful Titan): a new architecture with long-term in-context memory that learns how to
Mo Lotfollahi (@mo_lotfollahi) 's Twitter Profile Photo

امیدوارم که خانواده هاتون سالم باشن، چشمهامون اشک آلوده برای ایران، برای کودکان و ادم های بیگناه که به خاک و خون کشیده شدن. برای همه دانشجویان ایرانی در زمینه هوش مصنوعی و بیوانفرماتیک و زیست شناسی محاسباتی که دچار مشکل شدن برای ادامه تحصیل یا تحقیقاتشون، به من ایمیل بزننین برای

Erfan Loghmani (@loghmanierfan) 's Twitter Profile Photo

Can we align language models using observational data instead of costly experiments like A/B tests? In my latest research, we find that historical observational data *does* carry useful signals, but without causal care, models can learn the wrong things. Thread 🧵

Peymàn M. Kiasari (@pkiasari) 's Twitter Profile Photo

We’re presenting our work today at #ICML2025, Poster Session 4 West, W-214! If you are interested in computer vision reasoning and multimodal LLMs come visit us!

Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Great attending ICML this year, really enjoyed connecting with the graph learning community! Thanks a lot to Christopher Morris for organizing the lovely dinner :)

Hamed Shirzad (@hamedshirzad13) 's Twitter Profile Photo

Enjoyed giving our tutorial on Geometric & Topological Deep Learning at IEEE MLSP 2025 alongside Semih Cantürk. Loving the Istanbul vibes and the amazing food here! ✅

Enjoyed giving our tutorial on Geometric &amp; Topological Deep Learning at <a href="/ieee_mlsp/">IEEE MLSP</a> 2025 alongside <a href="/semihcanturk_en/">Semih Cantürk</a>. Loving the Istanbul vibes and the amazing food here! ✅