Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Foundation Models for Decision Making

Self-Select: Optimizing Instruction Selection for Large Language Models

Keshav Ramji · Alexander Kyimpopkin


Abstract: The same question can often be presented in different ways, depending on the audience and the intent with which it is being posed. To determine whether large language models (LLMs) demonstrate preferences for one phrasing over another regardless of semantic content, we introduce _Self-Select_, a method for selection of a preferred instruction template, and generation of high-quality synthetic data samples. This algorithm makes use of a _meta-prompt_ to decide on an instruction template, given the task and a list of candidate templates; then, the same LLM generates $n$ new samples following the structure of the chosen template. We evaluate _Self-Select_ on numerical reasoning and sentiment classification tasks, using a variety of both instruction-tuned and base models, providing insights into their ability to perform instruction selection, and their respective preferences and biases as a function of the instruction fine-tuning data. Our results demonstrate that stronger models can successfully perform instruction selection, and that large instruction-tuned models are more consistent and stable in their template choices. Furthermore, we find that permuting the instruction template ordering in the prompt leads to vastly different choice distributions, suggesting that decisions may be influenced more by inductive biases than by semantic understanding, even after instruction-tuning.

Chat is not available.