gal vardi (@galvardi) 's Twitter Profile
gal vardi

@galvardi

ID: 1539689263

calendar_today22-06-2013 22:55:06

10 Tweet

60 Followers

151 Following

Gilad Yehudai (@giladude) 's Twitter Profile Photo

Check out our latest work where we reconstruct large portions of the actual training data from trained neural networks. Joint work with gal vardi Niv Haim Project page: giladude1.github.io/reconstruction/ abs: arxiv.org/abs/2206.07758

Spencer Frei (@sfrei_) 's Twitter Profile Photo

New preprint with gal vardi, Peter Bartlett, Nati Srebro, and Wei Hu on the implicit bias of gradient descent/flow (GD/GF) in two-layer leaky ReLU networks when trained on high-dimensional data: arxiv.org/abs/2210.07082

Niv Haim (@haimniv) 's Twitter Profile Photo

Thrilled to share that our paper "๐˜™๐˜ฆ๐˜ค๐˜ฐ๐˜ฏ๐˜ด๐˜ต๐˜ณ๐˜ถ๐˜ค๐˜ต๐˜ช๐˜ฏ๐˜จ ๐˜›๐˜ณ๐˜ข๐˜ช๐˜ฏ๐˜ช๐˜ฏ๐˜จ ๐˜‹๐˜ข๐˜ต๐˜ข ๐˜ง๐˜ณ๐˜ฐ๐˜ฎ ๐˜›๐˜ณ๐˜ข๐˜ช๐˜ฏ๐˜ฆ๐˜ฅ ๐˜•๐˜ฆ๐˜ถ๐˜ณ๐˜ข๐˜ญ ๐˜•๐˜ฆ๐˜ต๐˜ธ๐˜ฐ๐˜ณ๐˜ฌ๐˜ด" was accepted to #NeurIPS 2022 as an ๐—ผ๐—ฟ๐—ฎ๐—น ๐—ฝ๐—ฟ๐—ฒ๐˜€๐—ฒ๐—ป๐˜๐—ฎ๐˜๐—ถ๐—ผ๐—ป! w/ Gilad Yehudai gal vardi Webpage: giladude1.github.io/reconstruction ๐Ÿงต1/n

Niv Haim (@haimniv) 's Twitter Profile Photo

Glad to share our new work: "Deconstructing Data Reconstruction: Multiclass, Weight Decay and General Losses" arxiv.org/abs/2307.01827 About trainset reconstruction in multiclass & regression tasks + perks (w/ Gon Buzaglo Gilad Yehudai gal vardi Yakir Oz Yaniv Nikankin M. Irani) ๐Ÿงต1/5

Glad to share our new work: "Deconstructing Data Reconstruction: Multiclass, Weight Decay and General Losses" arxiv.org/abs/2307.01827

About trainset reconstruction in multiclass & regression tasks + perks

(w/ <a href="/GonBuzaglo/">Gon Buzaglo</a> <a href="/Giladude/">Gilad Yehudai</a> <a href="/galvardi/">gal vardi</a>  <a href="/ozyakir/">Yakir Oz</a> <a href="/YNikankin/">Yaniv Nikankin</a> M. Irani)
๐Ÿงต1/5
Naftali Bennett ื ืคืชืœื™ ื‘ื ื˜ (@naftalibennett) 's Twitter Profile Photo

An entire family murdered in cold blood. Kedem family: Father Jonathan, mother Tamar, 6-year-old girls Shachar and Arbel, and 4-year-old boy Omer. Look at their happy faces. Their love. All of them murdered by Palestinian terrorists at Nir-oz kibbutz. Just because theyโ€™re

An entire family murdered in cold blood. 

Kedem family: Father Jonathan, mother Tamar, 6-year-old girls Shachar and Arbel, and 4-year-old boy Omer. 

Look at their happy faces. 
Their love. 

All of them murdered by Palestinian terrorists at Nir-oz kibbutz. 
Just because theyโ€™re
Ella Travels (Ella Kenan) (@ellatravelslove) 's Twitter Profile Photo

Tonight is the 4th night our babies will spend in Gaza. We do not know if they were harmed, injured, or even if they are still alive. This is the time to support Israel. Do everything you can to save our children, elders, men, and women from Hamas, a terrorist ISIS-like

Tonight is the 4th night our babies will spend in Gaza.  

We do not know if they were harmed, injured, or even if they are still alive. 

This is the time to support Israel. Do everything you can to save our children, elders, men, and women from Hamas, a terrorist ISIS-like
Spencer Frei (@sfrei_) 's Twitter Profile Photo

New preprint with gal vardi: arxiv.org/abs/2410.01774 We prove task- & sample-complexity bounds for in-context learning in transformers for classification tasks, focusing on a 1-layer linear attention architecture. We find they can exhibit benign overfitting in-context!

New preprint with <a href="/galvardi/">gal vardi</a>: arxiv.org/abs/2410.01774
We prove task- &amp; sample-complexity bounds for in-context learning in transformers for classification tasks, focusing on a 1-layer linear attention architecture. We find they can exhibit benign overfitting in-context!
Julia Kempe (@kempelab) 's Twitter Profile Photo

Optimization induces implicit bias. We study general steepest descent in homogeneous nets & show (generalized) convergence to a (generalized) KKT pt. Adam presents a curious case between l2 & l1: arxiv.org/abs/2410.22069 With Nikos Tsilivis & gal vardi NYU Center for Data Science AI at Meta

Optimization induces implicit bias. We study general steepest descent in homogeneous nets &amp; show (generalized) convergence to a (generalized) KKT pt. Adam presents a curious case between l2 &amp; l1: arxiv.org/abs/2410.22069
With Nikos Tsilivis &amp; <a href="/galvardi/">gal vardi</a> <a href="/NYUDataScience/">NYU Center for Data Science</a>  <a href="/AIatMeta/">AI at Meta</a>
Nirmit Joshi (@nirmitj_) 's Twitter Profile Photo

Happy to share our work on: "A Theory of Learning with Autoregressive Chain of Thought" arxiv.org/abs/2503.07932 Joint work with great collaborators: gal vardi, Adam Block, Surbhi Goel, Zhiyuan Li , Theodor Misiakiewicz, Nati Srebro