Cover Image for The Art of Evaluating Research Ideas w/ Chenglei Si
Cover Image for The Art of Evaluating Research Ideas w/ Chenglei Si
248 Went

The Art of Evaluating Research Ideas w/ Chenglei Si

Hosted by alphaXiv, Ron Arel & John Bohannon
Zoom
Registration
Past Event
Welcome! To join the event, please register below.
About Event

​​​🔬 AI4Science on alphaXiv – 4th Community Meetup
🗓 Wednesday July 9th 2025 · 10AM PT
🎙 Featuring Chenglei Si (Stanford NLP group)
💬 Casual Talk + Open Discussion

🎥 Zoom: https://stanford.zoom.us/j/96849944765?pwd=cPI5DaD9zEwV1TgFZV31O52ahhwsdp.1


Reliably evaluating research ideas is critical for identifying promising ideas to prioritize and to obtain reward signals for automated research agents. In this talk, Chenglei Si will discuss their two-year effort on evaluating LLM-generated ideas in comparison with human expert ideas, including both the evaluation of the ideas themselves as well as the evaluation of the executed projects. He will also discuss to what extent can LLM judges replace human expert reviewers in these evaluations.


​​​Whether you’re working on the frontier of LLMs or just curious about anything AI4Science, we’d love to have you there.

​​​Hosted by: alphaXiv x Intology 

​​​AI4Science: join the community

248 Went