diff --git a/docs/examples/agents/react.ipynb b/docs/examples/agents/react.ipynb index aef1239030f..7db37ddae72 100644 --- a/docs/examples/agents/react.ipynb +++ b/docs/examples/agents/react.ipynb @@ -38,7 +38,7 @@ }, { "cell_type": "code", - "execution_count": null, + "execution_count": 2, "id": "8078c8f1", "metadata": {}, "outputs": [ @@ -46,9 +46,34 @@ "name": "stdout", "output_type": "stream", "text": [ + "\n", + "\n", + "\u001b[1m> Entering new ReActDocstoreAgent chain...\u001b[0m\n", "Author David Chanoff has collaborated with a U.S. Navy admiral who served as the ambassador to the United Kingdom under which President?\n", - "Thought 1:" + "Thought 1:\u001b[32;1m\u001b[1;3m I need to search David Chanoff and the U.S. Navy admiral, find the ambassador to the United Kingdom, then find the President they served under.\n", + "Action 1: Search[David Chanoff]\u001b[0m\n", + "Observation 0: \u001b[36;1m\u001b[1;3mDavid Chanoff is a noted author of non-fiction work. His work has typically involved collaborations with the principal protagonist of the work concerned. His collaborators have included; Augustus A. White, Joycelyn Elders, Đoàn Văn Toại, William J. Crowe, Ariel Sharon, Kenneth Good and Felix Zandman. He has also written about a wide range of subjects including literary history, education and foreign for The Washington Post, The New Republic and The New York Times Magazine. He has published more than twelve books.\u001b[0m\n", + "Thought 1:\u001b[32;1m\u001b[1;3m David Chanoff has collaborated with many people, so I need to search the U.S. Navy admiral specifically.\n", + "Action 2: Search[U.S. Navy admiral]\u001b[0m\n", + "Observation 0: \u001b[36;1m\u001b[1;3mAdmiral of the Navy was the highest-possible rank in the United States Navy, prior to the creation of fleet admiral in 1944. The rank is considered to be at least equivalent to that of a five-star admiral, with Admiral George Dewey being the only officer to be appointed to the rank.\u001b[0m\n", + "Thought 1:\u001b[32;1m\u001b[1;3m I need to search the U.S. Navy admiral who served as the ambassador to the United Kingdom.\n", + "Action 3: Search[U.S. Navy admiral ambassador to the United Kingdom]\n", + "Observation 0: Admiral William J. Crowe Jr. was the United States Ambassador to the United Kingdom from 1994 to 1997. He served as Chairman of the Joint Chiefs of Staff from 1985 to 1989.\n", + "Thought 1: Admiral William J. Crowe Jr. served as the United States Ambassador to the United Kingdom from 1994 to 1997. So the President they served under is Bill Clinton.\n", + "Action 4: Finish[Bill Clinton]\n", + "Action 3: Finish[Bill Clinton]\u001b[0m\n", + "\u001b[1m> Finished ReActDocstoreAgent chain.\u001b[0m\n" ] + }, + { + "data": { + "text/plain": [ + "'Bill Clinton'" + ] + }, + "execution_count": 2, + "metadata": {}, + "output_type": "execute_result" } ], "source": [ diff --git a/langchain/agents/agent.py b/langchain/agents/agent.py index 80ee3ab1dc4..e2c2135b230 100644 --- a/langchain/agents/agent.py +++ b/langchain/agents/agent.py @@ -2,24 +2,19 @@ from __future__ import annotations from abc import ABC, abstractmethod -from typing import Any, ClassVar, Dict, List, NamedTuple, Optional, Tuple +from typing import Any, ClassVar, Dict, List, Optional, Tuple from pydantic import BaseModel from langchain.agents.tools import Tool from langchain.chains.base import Chain from langchain.chains.llm import LLMChain -from langchain.input import ChainedInput, get_color_mapping +from langchain.input import get_color_mapping +from langchain.agents.input import ChainedInput from langchain.llms.base import LLM from langchain.prompts.base import BasePromptTemplate - - -class Action(NamedTuple): - """Action to take.""" - - tool: str - tool_input: str - log: str +from langchain.schema import AgentAction +from langchain.logger import logger class Agent(Chain, BaseModel, ABC): @@ -99,7 +94,7 @@ class Agent(Chain, BaseModel, ABC): llm_chain = LLMChain(llm=llm, prompt=cls.create_prompt(tools)) return cls(llm_chain=llm_chain, tools=tools, **kwargs) - def get_action(self, text: str) -> Action: + def get_action(self, text: str) -> AgentAction: """Given input, decided what to do. Args: @@ -119,7 +114,7 @@ class Agent(Chain, BaseModel, ABC): full_output += output parsed_output = self._extract_tool_and_input(full_output) tool, tool_input = parsed_output - return Action(tool, tool_input, full_output) + return AgentAction(tool, tool_input, full_output) def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run text through and get agent response.""" @@ -135,7 +130,7 @@ class Agent(Chain, BaseModel, ABC): # prompts the LLM to take an action. starter_string = text + self.starter_string + self.llm_prefix # We use the ChainedInput class to iteratively add to the input over time. - chained_input = ChainedInput(starter_string, verbose=self.verbose) + chained_input = ChainedInput(starter_string, self.observation_prefix, self.llm_prefix, verbose=self.verbose) # We construct a mapping from each tool to a color, used for logging. color_mapping = get_color_mapping( [tool.name for tool in self.tools], excluded_colors=["green"] @@ -145,7 +140,7 @@ class Agent(Chain, BaseModel, ABC): # Call the LLM to see what to do. output = self.get_action(chained_input.input) # Add the log to the Chained Input. - chained_input.add(output.log, color="green") + chained_input.add_action(output, color="green") # If the tool chosen is the finishing tool, then we end and return. if output.tool == self.finish_tool_name: return {self.output_key: output.tool_input} @@ -154,8 +149,4 @@ class Agent(Chain, BaseModel, ABC): # We then call the tool on the tool input to get an observation observation = chain(output.tool_input) # We then log the observation - chained_input.add(f"\n{self.observation_prefix}") - chained_input.add(observation, color=color_mapping[output.tool]) - # We then add the LLM prefix into the prompt to get the LLM to start - # thinking, and start the loop all over. - chained_input.add(f"\n{self.llm_prefix}") + chained_input.add_observation(observation, color=color_mapping[output.tool]) diff --git a/langchain/agents/input.py b/langchain/agents/input.py new file mode 100644 index 00000000000..fec1dbe4b15 --- /dev/null +++ b/langchain/agents/input.py @@ -0,0 +1,40 @@ +"""Input manager for agents.""" +from typing import Optional + +from langchain.schema import AgentAction +from langchain.logger import logger + + +class ChainedInput: + """Class for working with input that is the result of chains.""" + + def __init__(self, text: str, observation_prefix: str, llm_prefix: str, verbose: bool = False): + """Initialize with verbose flag and initial text.""" + self._verbose = verbose + if self._verbose: + logger.log_agent_start(text) + self._input = text + self._observation_prefix = observation_prefix + self._llm_prefix = llm_prefix + + def add_action(self, action: AgentAction, color: Optional[str] = None) -> None: + """Add text to input, print if in verbose mode.""" + if self._verbose: + logger.log_agent_action(action, color=color) + self._input += action.log + + def add_observation(self, observation: str, color: Optional[str]) -> None: + """Add observation to input, print if in verbose mode.""" + if self._verbose: + logger.log_agent_observation( + observation, + color=color, + observation_prefix=self._observation_prefix, + llm_prefix=self._llm_prefix, + ) + self._input += f"\n{self._observation_prefix}{observation}\n{self._llm_prefix}" + + @property + def input(self) -> str: + """Return the accumulated input.""" + return self._input diff --git a/langchain/input.py b/langchain/input.py index 782cd41a8bf..680685fff51 100644 --- a/langchain/input.py +++ b/langchain/input.py @@ -27,25 +27,3 @@ def print_text(text: str, color: Optional[str] = None, end: str = "") -> None: else: color_str = _TEXT_COLOR_MAPPING[color] print(f"\u001b[{color_str}m\033[1;3m{text}\u001b[0m", end=end) - - -class ChainedInput: - """Class for working with input that is the result of chains.""" - - def __init__(self, text: str, verbose: bool = False): - """Initialize with verbose flag and initial text.""" - self._verbose = verbose - if self._verbose: - print_text(text, color=None) - self._input = text - - def add(self, text: str, color: Optional[str] = None) -> None: - """Add text to input, print if in verbose mode.""" - if self._verbose: - print_text(text, color=color) - self._input += text - - @property - def input(self) -> str: - """Return the accumulated input.""" - return self._input diff --git a/langchain/logger.py b/langchain/logger.py new file mode 100644 index 00000000000..546cda29ee9 --- /dev/null +++ b/langchain/logger.py @@ -0,0 +1,46 @@ +from langchain.schema import AgentAction +from typing import Optional, Any +from langchain.input import print_text +import logging +logging.basicConfig + + +class BaseLogger: + + def log_agent_start(self, text: str, **kwargs: Any): + pass + + def log_agent_end(self, text: str, **kwargs: Any): + pass + + def log_agent_action(self, action: AgentAction, **kwargs: Any): + pass + + def log_agent_observation(self, observation: str, **kwargs: Any): + pass + + +class StOutLogger(BaseLogger): + def log_agent_start(self, text: str, **kwargs: Any): + print_text(text) + + def log_agent_end(self, text: str, **kwargs: Any): + pass + + def log_agent_action(self, action: AgentAction, color: Optional[str] = None, **kwargs: Any): + print_text(action.log, color=color) + + def log_agent_observation( + self, + observation: str, + color: Optional[str] = None, + observation_prefix: Optional[str] = None, + llm_prefix: Optional[str] = None, + **kwargs: Any): + print_text(f"\n{observation_prefix}") + print_text(observation, color=color) + print_text(f"\n{llm_prefix}") + + + +logger = StOutLogger() \ No newline at end of file diff --git a/langchain/schema.py b/langchain/schema.py new file mode 100644 index 00000000000..389839f489c --- /dev/null +++ b/langchain/schema.py @@ -0,0 +1,11 @@ +from __future__ import annotations + +from typing import NamedTuple + + +class AgentAction(NamedTuple): + """Agent's action to take.""" + + tool: str + tool_input: str + log: str diff --git a/tests/unit_tests/test_input.py b/tests/unit_tests/test_input.py index 43dd3b080b8..dafc495520a 100644 --- a/tests/unit_tests/test_input.py +++ b/tests/unit_tests/test_input.py @@ -3,7 +3,8 @@ import sys from io import StringIO -from langchain.input import ChainedInput, get_color_mapping +from langchain.input import get_color_mapping +from langchain.agents.input import ChainedInput def test_chained_input_not_verbose() -> None: