This paper introduces ReAct, an approach that integrates reasoning and acting in large language models (LLMs). ReAct enables LLMs to generate both reasoning traces and task-specific actions in an interleaved manner. This synergy allows reasoning to help induce, track, and update action plans, while actions interface with external sources like knowledge bases to gather more information, overcoming issues of hallucination and error propagation in prior methods.
While large language models (LLMs) have shown impressive abilities, their capacities for reasoning (like chain-of-thought) and acting (like action plan generation) have been studied separately. This paper proposes ReAct, a new paradigm where LLMs generate both reasoning traces and task-specific actions in an interleaved fashion.
This approach creates a powerful synergy:
The ReAct framework was tested on a variety of language and decision-making tasks. In question answering (HotpotQA) and fact verification (Fever), it successfully mitigates issues of hallucination and error propagation common in chain-of-thought models. On interactive decision-making benchmarks like ALFWorld and WebShop, ReAct significantly outperformed imitation and reinforcement learning methods. The resulting task-solving trajectories are more human-like, interpretable, and trustworthy.