langchain/libs/partners/ollama/langchain_ollama
Elham Badri d696728278
partners/ollama: Enabled Token Level Streaming when Using Bind Tools for ChatOllama (#27689)
**Description:** The issue concerns the unexpected behavior observed
using the bind_tools method in LangChain's ChatOllama. When tools are
not bound, the llm.stream() method works as expected, returning
incremental chunks of content, which is crucial for real-time
applications such as conversational agents and live feedback systems.
However, when bind_tools([]) is used, the streaming behavior changes,
causing the output to be delivered in full chunks rather than
incrementally. This change negatively impacts the user experience by
breaking the real-time nature of the streaming mechanism.
**Issue:** #26971

---------

Co-authored-by: 4meyDam1e <amey.damle@mail.utoronto.ca>
Co-authored-by: Chester Curme <chester.curme@gmail.com>
2024-11-15 11:36:27 -05:00
..
__init__.py ollama: add pydocstyle linting for ollama (#27686) 2024-10-31 03:06:55 +00:00
chat_models.py partners/ollama: Enabled Token Level Streaming when Using Bind Tools for ChatOllama (#27689) 2024-11-15 11:36:27 -05:00
embeddings.py ollama: add pydocstyle linting for ollama (#27686) 2024-10-31 03:06:55 +00:00
llms.py
py.typed