James Thorne (@j6mes) 's Twitter Profile
James Thorne

@j6mes

CTO at Theia Insights. Assistant Prof at KAIST AI. PhD from @cambridge_cl. Co-organiser of fever workshop (fever.ai)

ID: 1950252162

linkhttps://jamesthorne.com calendar_today09-10-2013 23:19:59

763 Tweet

1,1K Followers

592 Following

James Thorne (@j6mes) 's Twitter Profile Photo

Pleasure to host Ohad Rubin at KAIST AI to hear about his ongoing work of retrieval augmented language models, long range modeling and RPT arxiv.org/abs/2306.13421

Pleasure to host <a href="/OhadRubin/">Ohad Rubin</a> at <a href="/kaist_ai/">KAIST AI</a> to hear about his ongoing work of retrieval augmented language models, long range modeling and RPT arxiv.org/abs/2306.13421
KAIST AI (@kaist_ai) 's Twitter Profile Photo

⭐️Congratulations to everyone at KAIST AI who have a total of 20 papers accepted to #NeurIPS2023! Check them here 🧵docs.google.com/document/d/1t5…

KAIST AI (@kaist_ai) 's Twitter Profile Photo

⭐️Congratulations to everyone at KAIST AI who has a total of 11 papers accepted to #EMNLP2023! Check them here 📷🧵 docs.google.com/document/d/1YB…

UKP Lab (@ukplab) 's Twitter Profile Photo

A group photo from the poster presentation of »AmbiFC: Fact-Checking Ambiguous Claims with Evidence«, co-authored by @Max_Glockner (UKP Lab), Ieva Staliūnaitė (Cambridge Computer Science), James Thorne (KAIST AI) , Gisela Vallejo (@unimelb), Andreas Vlachos (Cambridge Computer Science) and Iryna Gurevych. #EMNLP2023

A group photo from the poster presentation of »AmbiFC: Fact-Checking Ambiguous Claims with Evidence«, co-authored by @Max_Glockner (<a href="/UKPLab/">UKP Lab</a>), Ieva Staliūnaitė (<a href="/Cambridge_CL/">Cambridge Computer Science</a>), <a href="/j6mes/">James Thorne</a> (<a href="/kaist_ai/">KAIST AI</a>) , Gisela Vallejo (@unimelb), <a href="/vlachos_nlp/">Andreas Vlachos</a> (<a href="/Cambridge_CL/">Cambridge Computer Science</a>) and <a href="/IGurevych/">Iryna Gurevych</a>. #EMNLP2023
James Thorne (@j6mes) 's Twitter Profile Photo

Pleased to announce that Theia Insights has raised $6.5M to develop foundational AI technologies for the investment industry. Well done team!

Alice Oh (@aliceoh) 's Twitter Profile Photo

I can’t keep up with all these new papers and datasets from my students 😊 James Thorne and I wanted to give #kaist undergrad students research experience with LLMs in their own languages, and this is one of the outcomes of that project!

Daniel Vila Suero (@dvilasuero) 's Twitter Profile Photo

This is actually huge: - No SFT stage (e.g., Zephyr used 200k examples) - Preference tuning with 7K examples only (other models trained with at least 60k samples) I've put a lot of care & love building the DPO version of the amazing Capybara dataset from LDJ so I'm

This is actually huge: 

- No SFT stage (e.g., Zephyr used 200k examples)
- Preference tuning with 7K examples only (other models trained with at least 60k samples)

I've put a lot of care &amp; love building the DPO version of the amazing Capybara dataset from <a href="/ldjconfirmed/">LDJ</a> so I'm
Sayak Paul (@risingsayak) 's Twitter Profile Photo

Aligning a diffusion model on preference data WITHOUT a reference model could be nice no? So, Kashif Rasul and I are ideating the use of ORPO to align SDXL 1.0 on PickAPic. Diffusion ORPO with LoRA 💫 Code and model ⬇️ huggingface.co/sayakpaul/diff…

Aligning a diffusion model on preference data WITHOUT a reference model could be nice no?

So, <a href="/krasul/">Kashif Rasul</a> and I are ideating the use of ORPO to align SDXL 1.0 on PickAPic.

Diffusion ORPO with LoRA 💫

Code and model ⬇️ huggingface.co/sayakpaul/diff…
Alice Oh (@aliceoh) 's Twitter Profile Photo

Great work! I’m glad our benchmarks KoBBQ, KOLD, and CLIcK are used to evaluate HyperCLOVA_X. Making progress in Korean LLM! 🇰🇷🇰🇷🇰🇷 arxiv.org/abs/2307.16778 arxiv.org/abs/2205.11315 arxiv.org/abs/2403.06412

Jiwoo Hong @ NAACL 2025 (@jiwoohong98) 's Twitter Profile Photo

🔥Mixtral-8x22B-base + ORPO🔥 7k data & 1.3 hours to build a strong human-aligned 140B chat model🦾 👉IFEval: 65% 👉BBH: 59% 👉MT-Bench: 8.17 More models will be added to the Zephyr-ORPO collection with Argilla and Hugging Face , stay-tuned😃

Alice Oh (@aliceoh) 's Twitter Profile Photo

We are hosting wonderful NLP colleagues at KAIST on their way to ACL Bangkok! 🤩 On-site registration is closed, but the talks will be broadcast on Zoom. Please join us! Date/Time: Aug 10, 2024, 10:05-12:30 KST (UCT+9) Parallel Session 1: Advanced Language Models and AI

Valeria Ruscio (@rusciovaleria) 's Twitter Profile Photo

📢 Excited to share our new paper with Fabrizio Silvestri: "Beyond Position: How Rotary Embeddings Shape Representations and Memory in Autoregressive Transformers"! arxiv.org/abs/2410.18067 Keep reading to find out how RoPE affects Transformer models beyond just positional encoding 🧵

James Thorne (@j6mes) 's Twitter Profile Photo

NVIDIA's earnings call wasn’t about semiconductors — it was a vision for AI factories, 24/7 inference, and industry transformation. Using TIIC, we reveal $NVDA's diversification shows the future is broader than chips. What's your take? #AI #NVIDIA #LLMs linkedin.com/pulse/nvidias-…

James Thorne (@j6mes) 's Twitter Profile Photo

Looking forward to attending #NeurIPS2024 this week. Come meet Namgyu Ho who will present on block-transformer, which speeds up decoding by up to 20x.

Andreas Vlachos (@vlachos_nlp) 's Twitter Profile Photo

Pleased to announce the next FEVERworkshop at ACL2025! Regular workshop papers (ARR and direct submissions) due 15th of April! And new shared task focusing on reproducible and efficient verification of real world claims! Check fever.ai and get keen!

Jiwoo Hong @ NAACL 2025 (@jiwoohong98) 's Twitter Profile Photo

⁉️Why do reward models suffer from over-optimization in RLHF? We revisit how representations are learned during reward modeling, revealing “hidden state dispersion” as the key, with a simple fix! 🧵 Meet us at ICML Conference! 📅July 16th (Wed) 11AM–1:30PM 📍East Hall A-B E-2608

⁉️Why do reward models suffer from over-optimization in RLHF?

We revisit how representations are learned during reward modeling, revealing “hidden state dispersion” as the key, with a simple fix!
🧵

Meet us at <a href="/icmlconf/">ICML Conference</a>!
📅July 16th (Wed) 11AM–1:30PM
📍East Hall A-B E-2608