shiemannor (@shiemannor) 's Twitter Profile
shiemannor

@shiemannor

Prof@Technion, Researcher@Nvidia, Founder@Jether Energy. Trying to get machine learning to really work.

ID: 1077174278917885953

calendar_today24-12-2018 12:08:55

16 Tweet

275 Followers

12 Following

Guy Tennenholtz (@guytenn) 's Twitter Profile Photo

Check out our most recent work "On Covariate Shift of Latent Confounders in Imitation and Reinforcement Learning", where we talk about an important challenge of using expert data with hidden covariates. arxiv.org/pdf/2110.06539…

Check out our most recent work "On Covariate Shift of Latent Confounders in Imitation and Reinforcement Learning", where we talk about an important challenge of using expert data with hidden covariates. arxiv.org/pdf/2110.06539…
Chen Tessler (@chentessler) 's Twitter Profile Photo

1/ Excited to share that our latest work, Conditional Adversarial Latent Models [CALM], has been accepted to ACM SIGGRAPH 2023. 🧵👇 #reinforcementlearning #animation #games #isaacgym #siggraph2023 NVIDIA AI

1/ Excited to share that our latest work, Conditional Adversarial Latent Models [CALM], has been accepted to <a href="/siggraph/">ACM SIGGRAPH</a> 2023.

🧵👇

#reinforcementlearning #animation #games #isaacgym #siggraph2023 <a href="/NVIDIAAI/">NVIDIA AI</a>
Chen Tessler (@chentessler) 's Twitter Profile Photo

3/ What new capabilities does this unlock? Linearly interpolating between two motions (in the latent space) produces semantically meaningful transitions.

Chen Tessler (@chentessler) 's Twitter Profile Photo

4/ We can leverage the semantically meaningful latent space for high-level style-conditioned policies. For instance, a high-level policy tasked with moving in a specified direction can be urged to use a specified style via a latent-space similarity reward (here, crouching).

Chen Tessler (@chentessler) 's Twitter Profile Photo

5/ Combining the ability to control the style and direction a motion is produced -- we construct a finite state-machine to control the character, both the motion and direction it is performed.

Chen Tessler (@chentessler) 's Twitter Profile Photo

6/ This enables solving unseen tasks in various forms, without training -- overcoming the need for meticulous reward/termination design.

Chen Tessler (@chentessler) 's Twitter Profile Photo

7/ This work was made possible thanks to my amazing co-authors: Yoni Kasten, Yunrong Guo, Shie Mannor, Gal Chechik, and Jason Peng. More videos, a link to the paper, and code (coming real soon!): research.nvidia.com/labs/par/calm/

Gal Dalal (@dalalgal) 's Twitter Profile Photo

We released a multi-agent RL framework for network congestion control with the first public realistic network simulator! github.com/NVlabs/RLCC. Based on the amazing work of Benjamin Fuhrer and Chen Tessler

Aviv Tamar (@avivtamar1) 's Twitter Profile Photo

Want to learn / teach RL? Check out new book draft: Reinforcement Learning - Foundations sites.google.com/view/rlfoundat… W/ shiemannor and Yishay Mansour This is a rigorous first course in RL, based on our teaching at TAU CS and Technion ECE.

Want to learn / teach RL? 
Check out new book draft:
Reinforcement Learning - Foundations
sites.google.com/view/rlfoundat…
W/ <a href="/shiemannor/">shiemannor</a> and <a href="/YishayMansour/">Yishay Mansour</a>
This is a rigorous first course in RL, based on our teaching at TAU CS and Technion ECE.
Aviv Tamar (@avivtamar1) 's Twitter Profile Photo

For teachers, we also have a 40+ page exam booklet on our website. Why this book? There are several other excellent textbooks, including Sutton and Barto and Bertsekas and Tsitsiklis.

Aviv Tamar (@avivtamar1) 's Twitter Profile Photo

But for teaching RL, we wanted a book that is both rigorous (full proofs, analytical examples), covers what we feel is most relevant, and easy enough for undergrad teaching. The book is a focused one semester course for advanced undergrad/early grad covering key topics in depth.

Aviv Tamar (@avivtamar1) 's Twitter Profile Photo

We hope you find it useful! The book is still work in progress - we’d be grateful for comments, suggestions, omissions, and errors of any kind, at [email protected]

UriG (@uri_gadot) 's Twitter Profile Photo

Tired of manual #ComfyUI workflow design? While recent methods predict them, our new paper, FlowRL, introduces a Reinforcement Learning framework that learns to generate complex, novel workflows for you! paper [arxiv.org/abs/2505.21478]

Tired of manual #ComfyUI workflow design? While recent methods predict them, our new paper, FlowRL, introduces a Reinforcement Learning framework that learns to generate complex, novel workflows for you! 
paper [arxiv.org/abs/2505.21478]