Researchers have developed new methods for steering large language model (LLM) behaviors at inference time without sacrificing generation quality. One approach, Prompt-only SV (PrOSV), intervenes only on prompt tokens, outperforming traditional full-sequence steering vectors on benchmarks like AxBench. Another method, FLAS (Flow-based Activation Steering), learns a concept-conditioned velocity field to transport activations, consistently outperforming prompting on Gemma models. A third technique, SKOP (Steering via Key-Orthogonal Projections), constrains attention rerouting to preserve reasoning and retrieval performance, achieving a better trade-off between utility and steering efficacy. AI
Summary written by gemini-2.5-flash-lite from 5 sources. How we write summaries →
IMPACT New techniques for inference-time LLM control could enable more nuanced and reliable AI applications by improving steering accuracy and reducing performance degradation.
RANK_REASON Three new arXiv papers introduce novel methods for controlling LLM behavior at inference time.