
LLM4Eval Workshop
@llm4eval
The Third Workshop on Large Language Models for Evaluation in IR (LLM4IREval@SIGIR2025)
ID: 1759989664277299200
https://llm4eval.github.io/ 20-02-2024 17:13:52
41 Tweet
82 Followers
56 Following

Could large language models (LLMs) replace humans for relevance judgment tasks? #LLMJudge aims to explore this question by examining various factors, including the size of the LLM, the prompt design, and other relevant parameters. #LLMJudge LLM4Eval Workshop #LLMs #judgment #SIGIR


Great panel discussion with Ian Soboroff | [email protected], L. Dietz @[email protected] @[email protected], Charlie Clarke, Michael Ekstrand moderated by Bhaskar Mitra | ভাস্কর মিত্র at LLM4Eval Workshop. #SIGIR24


📢 Thrilled to announce that, "SynDL: A Large-Scale Synthetic Test Collection for Passage Retrieval" is publicly available. w/ Xi Wang, Emine Yilmaz, Nick Craswell, Bhaskar Mitra | ভাস্কর মিত্র, Paul Thomas Check it out: arxiv.org/abs/2408.16312




In today's 2nd Workshop on #LLMs for Evaluation in #InformationRetrieval (#LLM4Eval) at #WSDM2025 (09:15 CET), Edgar Meij, our Head of #AI Platforms Engineering, will deliver a keynote on "Synthetic Evaluations & #GenAI Application Development" bloom.bg/4hyocEi #AI


