Improving Reasoning Performance in Large Language Models via Representation Engineering

Publikation: Konference artikel i Proceeding eller bog/rapport kapitelKonferencebidrag i proceedingsForskningpeer review

Abstract

Recent advancements in large language models (LLMs) have resulted in increasingly anthropomorphic language concerning the ability of LLMs to reason. Whether reasoning in LLMs should be understood to be inherently different is, however, widely debated. We propose utilizing a representation engineering approach wherein model activations are read from the residual stream of an LLM when processing a reasoning task. The activations are used to derive a control vector that is applied to the model as an inference-time intervention, modulating the representational space of the model, to improve performance on the specified task. We publish the code for deriving control vectors and analyzing model representations. The method allows us to improve performance on reasoning benchmarks and assess how control vectors influence the final logit distribution of a model via metrics such as KL divergence and entropy. We apply control vectors to Mistral-7B-Instruct and a range of Pythia models on an inductive, a deductive and mathematical reasoning task. We show that an LLM can, to a certain degree, be controlled to improve its perceived reasoning ability by modulating activations. The intervention is dependent upon the ability to reliably extract the model's typical state when correctly solving a task. Our results suggest that reasoning performance can be modulated in the same manner as other information-processing tasks performed by LLMs and demonstrate that we are capable of improving performance on specific tasks via a simple intervention on the residual stream with no additional training.
OriginalsprogEngelsk
Titel13th International Conference on Learning Representations (ICLR 2025)
Antal sider18
Udgivelsessted Singapore
ForlagInternational Conference on Learning Representations (ICLR)
Publikationsdato22 jan. 2025
Sider1-18
ISBN (Elektronisk)9798331320850
StatusUdgivet - 22 jan. 2025
BegivenhedThe Thirteenth International Conference on Learning Representations - Singapore, Singapore
Varighed: 24 apr. 202528 apr. 2025
https://www.iclr.cc/Conferences/2025

Konference

KonferenceThe Thirteenth International Conference on Learning Representations
Land/OmrådeSingapore
BySingapore
Periode24/04/202528/04/2025
Internetadresse

Fingeraftryk

Dyk ned i forskningsemnerne om 'Improving Reasoning Performance in Large Language Models via Representation Engineering'. Sammen danner de et unikt fingeraftryk.

Citationsformater