community[minor]: Allow passing allow_dangerous_deserialization when loading LLM chain (#18894)

### Issue
Recently, the new `allow_dangerous_deserialization` flag was introduced
for preventing unsafe model deserialization that relies on pickle
without user's notice (#18696). Since then some LLMs like Databricks
requires passing in this flag with true to instantiate the model.

However, this breaks existing functionality to loading such LLMs within
a chain using `load_chain` method, because the underlying loader
function
[load_llm_from_config](f96dd57501/libs/langchain/langchain/chains/loading.py (L40))
 (and load_llm) ignores keyword arguments passed in. 

### Solution
This PR fixes this issue by propagating the
`allow_dangerous_deserialization` argument to the class loader iff the
LLM class has that field.

---------

Co-authored-by: Eugene Yurtsev <eyurtsev@gmail.com>
Co-authored-by: Bagatur <22008038+baskaryan@users.noreply.github.com>
This commit is contained in:
Yuki Watanabe
2024-03-27 00:07:55 +09:00
committed by GitHub
parent d7c14cb6f9
commit cfecbda48b
3 changed files with 130 additions and 64 deletions

View File

@@ -37,9 +37,9 @@ def _load_llm_chain(config: dict, **kwargs: Any) -> LLMChain:
"""Load LLM chain from config dict."""
if "llm" in config:
llm_config = config.pop("llm")
llm = load_llm_from_config(llm_config)
llm = load_llm_from_config(llm_config, **kwargs)
elif "llm_path" in config:
llm = load_llm(config.pop("llm_path"))
llm = load_llm(config.pop("llm_path"), **kwargs)
else:
raise ValueError("One of `llm` or `llm_path` must be present.")
@@ -59,9 +59,9 @@ def _load_hyde_chain(config: dict, **kwargs: Any) -> HypotheticalDocumentEmbedde
"""Load hypothetical document embedder chain from config dict."""
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
if "embeddings" in kwargs:
@@ -78,9 +78,9 @@ def _load_hyde_chain(config: dict, **kwargs: Any) -> HypotheticalDocumentEmbedde
def _load_stuff_documents_chain(config: dict, **kwargs: Any) -> StuffDocumentsChain:
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
@@ -107,9 +107,9 @@ def _load_map_reduce_documents_chain(
) -> MapReduceDocumentsChain:
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
@@ -118,12 +118,14 @@ def _load_map_reduce_documents_chain(
if "reduce_documents_chain" in config:
reduce_documents_chain = load_chain_from_config(
config.pop("reduce_documents_chain")
config.pop("reduce_documents_chain"), **kwargs
)
elif "reduce_documents_chain_path" in config:
reduce_documents_chain = load_chain(config.pop("reduce_documents_chain_path"))
reduce_documents_chain = load_chain(
config.pop("reduce_documents_chain_path"), **kwargs
)
else:
reduce_documents_chain = _load_reduce_documents_chain(config)
reduce_documents_chain = _load_reduce_documents_chain(config, **kwargs)
return MapReduceDocumentsChain(
llm_chain=llm_chain,
@@ -138,14 +140,22 @@ def _load_reduce_documents_chain(config: dict, **kwargs: Any) -> ReduceDocuments
if "combine_documents_chain" in config:
combine_document_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_document_chain_config)
combine_documents_chain = load_chain_from_config(
combine_document_chain_config, **kwargs
)
elif "combine_document_chain" in config:
combine_document_chain_config = config.pop("combine_document_chain")
combine_documents_chain = load_chain_from_config(combine_document_chain_config)
combine_documents_chain = load_chain_from_config(
combine_document_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
elif "combine_document_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_document_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_document_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -158,11 +168,11 @@ def _load_reduce_documents_chain(config: dict, **kwargs: Any) -> ReduceDocuments
collapse_documents_chain = None
else:
collapse_documents_chain = load_chain_from_config(
collapse_document_chain_config
collapse_document_chain_config, **kwargs
)
elif "collapse_documents_chain_path" in config:
collapse_documents_chain = load_chain(
config.pop("collapse_documents_chain_path")
config.pop("collapse_documents_chain_path"), **kwargs
)
elif "collapse_document_chain" in config:
collapse_document_chain_config = config.pop("collapse_document_chain")
@@ -170,11 +180,11 @@ def _load_reduce_documents_chain(config: dict, **kwargs: Any) -> ReduceDocuments
collapse_documents_chain = None
else:
collapse_documents_chain = load_chain_from_config(
collapse_document_chain_config
collapse_document_chain_config, **kwargs
)
elif "collapse_document_chain_path" in config:
collapse_documents_chain = load_chain(
config.pop("collapse_document_chain_path")
config.pop("collapse_document_chain_path"), **kwargs
)
return ReduceDocumentsChain(
@@ -190,17 +200,17 @@ def _load_llm_bash_chain(config: dict, **kwargs: Any) -> Any:
llm_chain = None
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
# llm attribute is deprecated in favor of llm_chain, here to support old configs
elif "llm" in config:
llm_config = config.pop("llm")
llm = load_llm_from_config(llm_config)
llm = load_llm_from_config(llm_config, **kwargs)
# llm_path attribute is deprecated in favor of llm_chain_path,
# its to support old configs
elif "llm_path" in config:
llm = load_llm(config.pop("llm_path"))
llm = load_llm(config.pop("llm_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
if "prompt" in config:
@@ -217,9 +227,9 @@ def _load_llm_bash_chain(config: dict, **kwargs: Any) -> Any:
def _load_llm_checker_chain(config: dict, **kwargs: Any) -> LLMCheckerChain:
if "llm" in config:
llm_config = config.pop("llm")
llm = load_llm_from_config(llm_config)
llm = load_llm_from_config(llm_config, **kwargs)
elif "llm_path" in config:
llm = load_llm(config.pop("llm_path"))
llm = load_llm(config.pop("llm_path"), **kwargs)
else:
raise ValueError("One of `llm` or `llm_path` must be present.")
if "create_draft_answer_prompt" in config:
@@ -264,17 +274,17 @@ def _load_llm_math_chain(config: dict, **kwargs: Any) -> LLMMathChain:
llm_chain = None
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
# llm attribute is deprecated in favor of llm_chain, here to support old configs
elif "llm" in config:
llm_config = config.pop("llm")
llm = load_llm_from_config(llm_config)
llm = load_llm_from_config(llm_config, **kwargs)
# llm_path attribute is deprecated in favor of llm_chain_path,
# its to support old configs
elif "llm_path" in config:
llm = load_llm(config.pop("llm_path"))
llm = load_llm(config.pop("llm_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
if "prompt" in config:
@@ -293,9 +303,9 @@ def _load_map_rerank_documents_chain(
) -> MapRerankDocumentsChain:
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
return MapRerankDocumentsChain(llm_chain=llm_chain, **config) # type: ignore[arg-type]
@@ -306,9 +316,9 @@ def _load_pal_chain(config: dict, **kwargs: Any) -> Any:
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
return PALChain(llm_chain=llm_chain, **config) # type: ignore[arg-type]
@@ -317,18 +327,18 @@ def _load_pal_chain(config: dict, **kwargs: Any) -> Any:
def _load_refine_documents_chain(config: dict, **kwargs: Any) -> RefineDocumentsChain:
if "initial_llm_chain" in config:
initial_llm_chain_config = config.pop("initial_llm_chain")
initial_llm_chain = load_chain_from_config(initial_llm_chain_config)
initial_llm_chain = load_chain_from_config(initial_llm_chain_config, **kwargs)
elif "initial_llm_chain_path" in config:
initial_llm_chain = load_chain(config.pop("initial_llm_chain_path"))
initial_llm_chain = load_chain(config.pop("initial_llm_chain_path"), **kwargs)
else:
raise ValueError(
"One of `initial_llm_chain` or `initial_llm_chain_path` must be present."
)
if "refine_llm_chain" in config:
refine_llm_chain_config = config.pop("refine_llm_chain")
refine_llm_chain = load_chain_from_config(refine_llm_chain_config)
refine_llm_chain = load_chain_from_config(refine_llm_chain_config, **kwargs)
elif "refine_llm_chain_path" in config:
refine_llm_chain = load_chain(config.pop("refine_llm_chain_path"))
refine_llm_chain = load_chain(config.pop("refine_llm_chain_path"), **kwargs)
else:
raise ValueError(
"One of `refine_llm_chain` or `refine_llm_chain_path` must be present."
@@ -349,9 +359,13 @@ def _load_refine_documents_chain(config: dict, **kwargs: Any) -> RefineDocuments
def _load_qa_with_sources_chain(config: dict, **kwargs: Any) -> QAWithSourcesChain:
if "combine_documents_chain" in config:
combine_documents_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_documents_chain_config)
combine_documents_chain = load_chain_from_config(
combine_documents_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -369,13 +383,13 @@ def _load_sql_database_chain(config: dict, **kwargs: Any) -> Any:
raise ValueError("`database` must be present.")
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
chain = load_chain_from_config(llm_chain_config)
return SQLDatabaseChain(llm_chain=chain, database=database, **config) # type: ignore[arg-type]
chain = load_chain_from_config(llm_chain_config, **kwargs, **kwargs)
return SQLDatabaseChain(llm_chain=chain, database=database, **config)
if "llm" in config:
llm_config = config.pop("llm")
llm = load_llm_from_config(llm_config)
llm = load_llm_from_config(llm_config, **kwargs)
elif "llm_path" in config:
llm = load_llm(config.pop("llm_path"))
llm = load_llm(config.pop("llm_path"), **kwargs)
else:
raise ValueError("One of `llm` or `llm_path` must be present.")
if "prompt" in config:
@@ -396,9 +410,13 @@ def _load_vector_db_qa_with_sources_chain(
raise ValueError("`vectorstore` must be present.")
if "combine_documents_chain" in config:
combine_documents_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_documents_chain_config)
combine_documents_chain = load_chain_from_config(
combine_documents_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -418,9 +436,13 @@ def _load_retrieval_qa(config: dict, **kwargs: Any) -> RetrievalQA:
raise ValueError("`retriever` must be present.")
if "combine_documents_chain" in config:
combine_documents_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_documents_chain_config)
combine_documents_chain = load_chain_from_config(
combine_documents_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -442,9 +464,13 @@ def _load_retrieval_qa_with_sources_chain(
raise ValueError("`retriever` must be present.")
if "combine_documents_chain" in config:
combine_documents_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_documents_chain_config)
combine_documents_chain = load_chain_from_config(
combine_documents_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -464,9 +490,13 @@ def _load_vector_db_qa(config: dict, **kwargs: Any) -> VectorDBQA:
raise ValueError("`vectorstore` must be present.")
if "combine_documents_chain" in config:
combine_documents_chain_config = config.pop("combine_documents_chain")
combine_documents_chain = load_chain_from_config(combine_documents_chain_config)
combine_documents_chain = load_chain_from_config(
combine_documents_chain_config, **kwargs
)
elif "combine_documents_chain_path" in config:
combine_documents_chain = load_chain(config.pop("combine_documents_chain_path"))
combine_documents_chain = load_chain(
config.pop("combine_documents_chain_path"), **kwargs
)
else:
raise ValueError(
"One of `combine_documents_chain` or "
@@ -486,12 +516,14 @@ def _load_graph_cypher_chain(config: dict, **kwargs: Any) -> GraphCypherQAChain:
raise ValueError("`graph` must be present.")
if "cypher_generation_chain" in config:
cypher_generation_chain_config = config.pop("cypher_generation_chain")
cypher_generation_chain = load_chain_from_config(cypher_generation_chain_config)
cypher_generation_chain = load_chain_from_config(
cypher_generation_chain_config, **kwargs
)
else:
raise ValueError("`cypher_generation_chain` must be present.")
if "qa_chain" in config:
qa_chain_config = config.pop("qa_chain")
qa_chain = load_chain_from_config(qa_chain_config)
qa_chain = load_chain_from_config(qa_chain_config, **kwargs)
else:
raise ValueError("`qa_chain` must be present.")
@@ -506,7 +538,7 @@ def _load_graph_cypher_chain(config: dict, **kwargs: Any) -> GraphCypherQAChain:
def _load_api_chain(config: dict, **kwargs: Any) -> APIChain:
if "api_request_chain" in config:
api_request_chain_config = config.pop("api_request_chain")
api_request_chain = load_chain_from_config(api_request_chain_config)
api_request_chain = load_chain_from_config(api_request_chain_config, **kwargs)
elif "api_request_chain_path" in config:
api_request_chain = load_chain(config.pop("api_request_chain_path"))
else:
@@ -515,9 +547,9 @@ def _load_api_chain(config: dict, **kwargs: Any) -> APIChain:
)
if "api_answer_chain" in config:
api_answer_chain_config = config.pop("api_answer_chain")
api_answer_chain = load_chain_from_config(api_answer_chain_config)
api_answer_chain = load_chain_from_config(api_answer_chain_config, **kwargs)
elif "api_answer_chain_path" in config:
api_answer_chain = load_chain(config.pop("api_answer_chain_path"))
api_answer_chain = load_chain(config.pop("api_answer_chain_path"), **kwargs)
else:
raise ValueError(
"One of `api_answer_chain` or `api_answer_chain_path` must be present."
@@ -537,9 +569,9 @@ def _load_api_chain(config: dict, **kwargs: Any) -> APIChain:
def _load_llm_requests_chain(config: dict, **kwargs: Any) -> LLMRequestsChain:
if "llm_chain" in config:
llm_chain_config = config.pop("llm_chain")
llm_chain = load_chain_from_config(llm_chain_config)
llm_chain = load_chain_from_config(llm_chain_config, **kwargs)
elif "llm_chain_path" in config:
llm_chain = load_chain(config.pop("llm_chain_path"))
llm_chain = load_chain(config.pop("llm_chain_path"), **kwargs)
else:
raise ValueError("One of `llm_chain` or `llm_chain_path` must be present.")
if "requests_wrapper" in kwargs: