From 54e9ea433a1cdfee031819d4c9ad1f97608030ad Mon Sep 17 00:00:00 2001 From: gbaian10 <34255899+gbaian10@users.noreply.github.com> Date: Fri, 2 Aug 2024 23:32:56 +0800 Subject: [PATCH] fix: Modify the order of init_chat_model import ollama package. (#24977) --- libs/langchain/langchain/chat_models/base.py | 18 ++++++++---------- 1 file changed, 8 insertions(+), 10 deletions(-) diff --git a/libs/langchain/langchain/chat_models/base.py b/libs/langchain/langchain/chat_models/base.py index b98ef99ce0b..ce50f7b7b09 100644 --- a/libs/langchain/langchain/chat_models/base.py +++ b/libs/langchain/langchain/chat_models/base.py @@ -347,16 +347,14 @@ def _init_chat_model_helper( _check_pkg("langchain_ollama") from langchain_ollama import ChatOllama except ImportError: - pass - - # For backwards compatibility - try: - _check_pkg("langchain_community") - from langchain_community.chat_models import ChatOllama - except ImportError: - # If both langchain-ollama and langchain-community aren't available, raise - # an error related to langchain-ollama - _check_pkg("langchain_ollama") + # For backwards compatibility + try: + _check_pkg("langchain_community") + from langchain_community.chat_models import ChatOllama + except ImportError: + # If both langchain-ollama and langchain-community aren't available, + # raise an error related to langchain-ollama + _check_pkg("langchain_ollama") return ChatOllama(model=model, **kwargs) elif model_provider == "together":