feat: Add LangSmith integration metadata to create_agent and init_chat_model (#35810)

This commit is contained in:
Tanushree
2026-03-18 08:24:08 -07:00
committed by GitHub
parent dd136337d7
commit 2319fdc978
4 changed files with 40 additions and 11 deletions

View File

@@ -69,6 +69,8 @@ class LangSmithParams(TypedDict, total=False):
ls_stop: list[str] | None
"""Stop words for generation."""
ls_integration: str
"""Integration that created the trace."""
@cache # Cache the tokenizer
@@ -299,6 +301,22 @@ class BaseLanguageModel(
# generate responses that match a given schema.
raise NotImplementedError
def _get_ls_params(
self,
stop: list[str] | None = None, # noqa: ARG002
**kwargs: Any, # noqa: ARG002
) -> LangSmithParams:
"""Get standard params for tracing."""
return LangSmithParams()
def _get_ls_params_with_defaults(
self,
stop: list[str] | None = None,
**kwargs: Any,
) -> LangSmithParams:
"""Wrap _get_ls_params to include any additional default parameters."""
return self._get_ls_params(stop=stop, **kwargs)
@property
def _identifying_params(self) -> Mapping[str, Any]:
"""Get the identifying parameters."""

View File

@@ -505,7 +505,7 @@ class BaseChatModel(BaseLanguageModel[AIMessage], ABC):
options = {"stop": stop, **kwargs, **ls_structured_output_format_dict}
inheritable_metadata = {
**(config.get("metadata") or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = CallbackManager.configure(
config.get("callbacks"),
@@ -633,7 +633,7 @@ class BaseChatModel(BaseLanguageModel[AIMessage], ABC):
options = {"stop": stop, **kwargs, **ls_structured_output_format_dict}
inheritable_metadata = {
**(config.get("metadata") or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = AsyncCallbackManager.configure(
config.get("callbacks"),
@@ -827,6 +827,16 @@ class BaseChatModel(BaseLanguageModel[AIMessage], ABC):
return ls_params
def _get_ls_params_with_defaults(
self,
stop: list[str] | None = None,
**kwargs: Any,
) -> LangSmithParams:
"""Wrap _get_ls_params to always include ls_integration."""
ls_params = self._get_ls_params(stop=stop, **kwargs)
ls_params["ls_integration"] = "langchain_chat_model"
return ls_params
def _get_llm_string(self, stop: list[str] | None = None, **kwargs: Any) -> str:
if self.is_lc_serializable():
params = {**kwargs, "stop": stop}
@@ -899,7 +909,7 @@ class BaseChatModel(BaseLanguageModel[AIMessage], ABC):
options = {"stop": stop, **ls_structured_output_format_dict}
inheritable_metadata = {
**(metadata or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = CallbackManager.configure(
@@ -1022,7 +1032,7 @@ class BaseChatModel(BaseLanguageModel[AIMessage], ABC):
options = {"stop": stop, **ls_structured_output_format_dict}
inheritable_metadata = {
**(metadata or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = AsyncCallbackManager.configure(

View File

@@ -527,7 +527,7 @@ class BaseLLM(BaseLanguageModel[str], ABC):
options = {"stop": stop}
inheritable_metadata = {
**(config.get("metadata") or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = CallbackManager.configure(
config.get("callbacks"),
@@ -597,7 +597,7 @@ class BaseLLM(BaseLanguageModel[str], ABC):
options = {"stop": stop}
inheritable_metadata = {
**(config.get("metadata") or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
callback_manager = AsyncCallbackManager.configure(
config.get("callbacks"),
@@ -906,14 +906,14 @@ class BaseLLM(BaseLanguageModel[str], ABC):
metadata = [
{
**(meta or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
for meta in metadata
]
elif isinstance(metadata, dict):
metadata = {
**(metadata or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
if (
isinstance(callbacks, list)
@@ -1173,14 +1173,14 @@ class BaseLLM(BaseLanguageModel[str], ABC):
metadata = [
{
**(meta or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
for meta in metadata
]
elif isinstance(metadata, dict):
metadata = {
**(metadata or {}),
**self._get_ls_params(stop=stop, **kwargs),
**self._get_ls_params_with_defaults(stop=stop, **kwargs),
}
# Create callback managers
if isinstance(callbacks, list) and (

View File

@@ -1630,8 +1630,9 @@ def create_agent(
)
config: RunnableConfig = {"recursion_limit": 10_000}
config["metadata"] = {"ls_integration": "langchain_create_agent"}
if name:
config["metadata"] = {"lc_agent_name": name}
config["metadata"]["lc_agent_name"] = name
return graph.compile(
checkpointer=checkpointer,