Precision Shapes Personality: The Hidden Cost of Quantization in Sub-Billion-LLMs
Abstract
Psychometric studies of language models are increasingly important given their growing use as human assistants and in therapeutic settings. Such applications are often deployed on edge devices with sub-billion parameter large language models (LLMs) operating under strict memory and latency constraints, where post-training quantization (PTQ) is standard. Yet little is known about whether numeric precision alters measured personality traits. In the current work, using a psychometric benchmark TRAIT, we evaluate five sub-1B LLMs across different precision settings. We find that 4-bit Normal Float (nf4) produces the largest shifts, int8 smaller ones, and 16-bit formats remain closest to native. Shifts concentrate in Extraversion, Conscientiousness, and Narcissism, while Openness and Machiavellianism are more stable. These results identify precision as a consequential, controllable variable that should be disclosed and audited when personality matters in deployment.