langchain/libs/partners/ollama/langchain_ollama
Elham Badri d696728278
partners/ollama: Enabled Token Level Streaming when Using Bind Tools for ChatOllama (#27689)
**Description:** The issue concerns the unexpected behavior observed
using the bind_tools method in LangChain's ChatOllama. When tools are
not bound, the llm.stream() method works as expected, returning
incremental chunks of content, which is crucial for real-time
applications such as conversational agents and live feedback systems.
However, when bind_tools([]) is used, the streaming behavior changes,
causing the output to be delivered in full chunks rather than
incrementally. This change negatively impacts the user experience by
breaking the real-time nature of the streaming mechanism.
**Issue:** #26971

---------

Co-authored-by: 4meyDam1e <amey.damle@mail.utoronto.ca>
Co-authored-by: Chester Curme <chester.curme@gmail.com>
2024-11-15 11:36:27 -05:00
..
__init__.py ollama: add pydocstyle linting for ollama (#27686) 2024-10-31 03:06:55 +00:00
chat_models.py partners/ollama: Enabled Token Level Streaming when Using Bind Tools for ChatOllama (#27689) 2024-11-15 11:36:27 -05:00
embeddings.py ollama: add pydocstyle linting for ollama (#27686) 2024-10-31 03:06:55 +00:00
llms.py multiple: pydantic 2 compatibility, v0.3 (#26443) 2024-09-13 14:38:45 -07:00
py.typed ollama: init package (#23615) 2024-07-20 00:43:29 +00:00