firstbacksecondback
23 Results
Workshop
|
Evaluating Gender Bias Transfer between Pre-trained and Prompt Adapted Language Models Nivedha Sivakumar · Natalie Mackraz · Samira Khorshidi · Krishna Patel · Barry-John Theobald · Luca Zappella · Nicholas Apostoloff |
||
Workshop
|
Sat 15:45 |
Statistical Uncertainty Quantification for Aggregate Performance Metrics in Machine Learning Benchmarks Rachel Longjohn · Giri Gopalan · Emily Casleton |
|
Workshop
|
Improving Bias Metrics in Vision-Language Models by Addressing Inherent Model Disabilities Lakshmipathi Balaji Darur · Shanmukha Sai Keerthi Gouravarapu · Shashwat Goel · Ponnurangam Kumaraguru |
||
Workshop
|
Sat 17:27 |
Improving Bias Metrics in Vision-Language Models by Addressing Inherent Model Disabilities Lakshmipathi Balaji Darur · Shanmukha Sai Keerthi Gouravarapu · Shashwat Goel · Ponnurangam Kumaraguru |
|
Workshop
|
Different Bias Under Different Criteria: Assessing Bias in LLMs with a Fact-Based Approach Changgeon Ko · Jisu Shin · Hoyun Song · Jeongyeon Seo · Jong Park |
||
Workshop
|
Understanding Memorization using Representation Similarity Analysis and Model Stitching Aishwarya Gupta · Indranil Saha · Piyush Rai |
||
Workshop
|
MED-OMIT: Extrinsically-Focused Evaluation Metric for Omissions in Medical Summarization Elliot Schumacher · Daniel Rosenthal · Dhruv Naik · Varun Nair · Luladay Price · Geoffrey Tso · Anitha Kannan |
||
Workshop
|
Decoding Biases: An Analysis of Automated Methods and Metrics for Gender Bias Detection in Language Models Shachi H. Kumar · Saurav Sahay · Sahisnu Mazumder · Eda Okur · Ramesh Manuvinakurike · Nicole Beckage · Hsuan Su · Hung-yi Lee · Lama Nachman |
||
Workshop
|
Sat 16:15 |
Benchmarking Music Generation Models and Metrics via Human Preference Studies Ahmet Solak · Florian Grötschla · Luca Lanzendörfer · Roger Wattenhofer |
|
Workshop
|
Assessing Bias in Metric Models for LLM Open-Ended Generation Bias Benchmarks Nathaniel Demchak · Xin Guan · Zekun Wu · Ziyi Xu · Adriano Koshiyama · Emre Kazim |
||
Workshop
|
Sat 17:27 |
Toward Large Language Models that Benefit for All: Benchmarking Group Fairness in Reward Models Kefan Song · Jin Yao · Shangtong Zhang |
|
Workshop
|
Large Language Model Detoxification: Data and Metric Solutions SungJoo Byun · HYOPIL SHIN |