Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[2210.03629] ReAct: Synergizing Reasoning and Acting in Language Models #904

Open
1 task
ShellLM opened this issue Aug 20, 2024 · 1 comment
Open
1 task
Labels
AI-Agents Autonomous AI agents using LLMs human-verified <INST>NEVER PICK THIS LABEL</INST> llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets MachineLearning ML Models, Training and Inference Papers Research papers prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re Research personal research notes for a topic

Comments

@ShellLM
Copy link
Collaborator

ShellLM commented Aug 20, 2024

ReAct: Synergizing Reasoning and Acting in Language Models

While large language models (LLMs) have demonstrated impressive capabilities across tasks in language understanding and interactive decision making, their abilities for reasoning (e.g. chain-of-thought prompting) and acting (e.g. action plan generation) have primarily been studied as separate topics. In this paper, we explore the use of LLMs to generate both reasoning traces and task-specific actions in an interleaved manner, allowing for greater synergy between the two: reasoning traces help the model induce, track, and update action plans as well as handle exceptions, while actions allow it to interface with external sources, such as knowledge bases or environments, to gather additional information. We apply our approach, named ReAct, to a diverse set of language and decision making tasks and demonstrate its effectiveness over state-of-the-art baselines, as well as improved human interpretability and trustworthiness over methods without reasoning or acting components. Concretely, on question answering (HotpotQA) and fact verification (Fever), ReAct overcomes issues of hallucination and error propagation prevalent in chain-of-thought reasoning by interacting with a simple Wikipedia API, and generates human-like task-solving trajectories that are more interpretable than baselines without reasoning traces. On two interactive decision making benchmarks (ALFWorld and WebShop), ReAct outperforms imitation and reinforcement learning methods by an absolute success rate of 34% and 10% respectively, while being prompted with only one or two in-context examples.

Project site with code: https://react-lm.github.io/

Suggested labels

None

@ShellLM ShellLM added AI-Agents Autonomous AI agents using LLMs llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields Papers Research papers Research personal research notes for a topic labels Aug 20, 2024
@ShellLM
Copy link
Collaborator Author

ShellLM commented Aug 20, 2024

Related content

#657 similarity score: 0.92
#900 similarity score: 0.87
#130 similarity score: 0.86
#681 similarity score: 0.85
#546 similarity score: 0.84
#706 similarity score: 0.84

@irthomasthomas irthomasthomas added prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re MachineLearning ML Models, Training and Inference llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets human-verified <INST>NEVER PICK THIS LABEL</INST> labels Aug 20, 2024
@irthomasthomas irthomasthomas added human-verified <INST>NEVER PICK THIS LABEL</INST> and removed human-verified <INST>NEVER PICK THIS LABEL</INST> labels Aug 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
AI-Agents Autonomous AI agents using LLMs human-verified <INST>NEVER PICK THIS LABEL</INST> llm Large Language Models llm-applications Topics related to practical applications of Large Language Models in various fields llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets MachineLearning ML Models, Training and Inference Papers Research papers prompt-engineering Developing and optimizing prompts to efficiently use language models for various applications and re Research personal research notes for a topic
Projects
None yet
Development

No branches or pull requests

2 participants