Rasika Bhalerao (@rasikabh) 's Twitter Profile
Rasika Bhalerao

@rasikabh

I teach CS at Northeastern University. She/her

ID: 3234618774

linkhttps://rasikabh.github.io/ calendar_today03-06-2015 03:47:46

14 Tweet

84 Followers

152 Following

NYU Center for Data Science (@nyudatascience) 's Twitter Profile Photo

CDS Faculty Sam Bowman recently published his research on how our models learn bias. Read about the press coverage for the paper as well as Professor Bowman’s thoughts about the research in our latest blog post: medium.com/@NYUDataScienc…

New York Times World (@nytimesworld) 's Twitter Profile Photo

Margo St. James, who sought to decriminalize prostitution and make life better for sex workers, has died at 83. An erstwhile sex worker herself, she would begin speeches by saying, "Nice to see so many familiar faces." Obit by Katharine Q. Seelye nyti.ms/3itnZpg

SWE Grad Community (@swe_grad) 's Twitter Profile Photo

Blog posts on mindfulness in grad school (gradswe.swe.org/gradswe-blog/m…), support from software in grad school (gradswe.swe.org/gradswe-blog/s…), and a pep talk for members who are several years into a long degree (gradswe.swe.org/gradswe-blog/a…)!

Elissa M. Redmiles, Ph.D. (@eredmil1) 's Twitter Profile Photo

Are you an #onlyfans creator? Sign up for our paid ($50/up to 75 min) interview study. We want to understand your digital experiences! bit.ly/OnlyFansStudy

Are you an #onlyfans creator? Sign up for our paid ($50/up to 75 min) interview study.

We want to understand your digital experiences!

bit.ly/OnlyFansStudy
Tomek Korbak (@tomekkorbak) 's Twitter Profile Photo

You can (and should) do RL from human feedback during pretraining itself! In our new paper, we show how training w/ human preferences early on greatly reduces undesirable LM behaviors, including under adversarial attack, w/o hurting downstream performance. arxiv.org/abs/2302.08582

You can (and should) do RL from human feedback during pretraining itself! In our new paper, we show how training w/ human preferences early on greatly reduces undesirable LM behaviors, including under adversarial attack, w/o hurting downstream performance. arxiv.org/abs/2302.08582
Tomek Korbak (@tomekkorbak) 's Twitter Profile Photo

Excited to announce our paper Pretraining Language Models with Human Preferences has been accepted as an oral presentation at #ICML2023! 🚀