Beyond Labels: Explanatory Collapse due to Instruction Tuning in Protein LLMs
Yining Yang · Ruihong Huang · Yang Shen
Abstract
Instruction tuning on domain-specific datasets improves categorical accuracy but consistently degrades explanatory behavior. In protein large language models, we observe that tuned models ignore explanation requests, fall back on generic templates, and produce degenerate structured outputs with trivial repetitions or hallucinated lists. This collapse of expressive diversity renders models terse and uninformative, limiting their scientific utility. Our findings highlight a trade-off in current protein instruction-tuning practices: accuracy is gained at the cost of interpretive value, underscoring the need for strategies that preserve explanatory depth.
Video
Chat is not available.
Successful Page Load