Workshop
|
|
Q-Morality: Quantum-Enhanced ActAdd-Guided Bias Reduction in LLMs
Shardul Kulkarni
|
|
Oral
|
Thu 10:20
|
LLM Evaluators Recognize and Favor Their Own Generations
Arjun Panickssery · Samuel Bowman · Shi Feng
|
|
Workshop
|
Sat 12:00
|
A STEP TOWARDS MIXTURE OF GRADER: STATISTICAL ANALYSIS OF EXISTING AUTOMATIC EVALUATION METRICS
Yun Joon Soh · Jishen Zhao
|
|
Poster
|
Thu 11:00
|
LLM Evaluators Recognize and Favor Their Own Generations
Arjun Panickssery · Samuel Bowman · Shi Feng
|
|
Workshop
|
|
Moral Persuasion in Large Language Models: Evaluating Susceptibility and Ethical Alignment
Allison Huang · Carlos Mougan · Yulu Pi
|
|
Workshop
|
|
Rethinking CyberSecEval: An LLM-Aided Approach to Evaluation Critique
Suhas Hariharan · Zainab Ali Majid · Jaime Raldua Veuthey · Jacob Haimes
|
|
Workshop
|
|
THaMES: An End-to-End Tool for Hallucination Mitigation and Evaluation in Large Language Models
Mengfei Liang · Archish Arun · Zekun Wu · CRISTIAN VILLALOBOS · Jonathan Lutch · Emre Kazim · Adriano Koshiyama · Philip Treleaven
|
|
Workshop
|
|
Analyzing Probabilistic Methods for Evaluating Agent Capabilities
Axel Højmark · Govind Pimpale · Arjun Panickssery · Marius Hobbhahn · Jérémy Scheurer
|
|
Workshop
|
|
Towards Better Fairness Metrics for Counter-Human Trafficking AI Initiatives
Vidya Sujaya · Pratheeksha Nair · Reihaneh Rabbany
|
|
Workshop
|
|
Understanding The Effect Of Temperature On Alignment With Human Opinions
Maja Pavlovic · Massimo Poesio
|
|
Workshop
|
Sat 17:27
|
Understanding The Effect Of Temperature On Alignment With Human Opinions
Maja Pavlovic · Massimo Poesio
|
|
Workshop
|
|
Code-Switching Red-Teaming: LLM Evaluation for Safety and Multilingual Understanding
Haneul Yoo · Yongjin Yang · Hwaran Lee
|
|