mirror of
https://github.com/hwchase17/langchain.git
synced 2025-08-12 22:28:03 +00:00
These are set in Github workflows, but forgot to add them to most makefiles for convenience when developing locally. `uv run` will automatically sync the lock file. Because many of our development dependencies are local installs, it will pick up version changes and update the lock file. Passing `--frozen` or setting this environment variable disables the behavior. |
||
---|---|---|
.. | ||
langchain_ollama | ||
scripts | ||
tests | ||
.gitignore | ||
LICENSE | ||
Makefile | ||
pyproject.toml | ||
README.md | ||
uv.lock |
langchain-ollama
This package contains the LangChain integration with Ollama
Installation
pip install -U langchain-ollama
You will also need to run the Ollama server locally. You can download it here.
Chat Models
ChatOllama
class exposes chat models from Ollama.
from langchain_ollama import ChatOllama
llm = ChatOllama(model="llama3-groq-tool-use")
llm.invoke("Sing a ballad of LangChain.")
Embeddings
OllamaEmbeddings
class exposes embeddings from Ollama.
from langchain_ollama import OllamaEmbeddings
embeddings = OllamaEmbeddings(model="llama3")
embeddings.embed_query("What is the meaning of life?")
LLMs
OllamaLLM
class exposes LLMs from Ollama.
from langchain_ollama import OllamaLLM
llm = OllamaLLM(model="llama3")
llm.invoke("The meaning of life is")