mirror of
https://github.com/hwchase17/langchain.git
synced 2025-07-05 04:38:26 +00:00
docs: Hugging Face
update (#16490)
- added missed integrations to the platform page - updated integration examples: added links and fixed formats
This commit is contained in:
parent
c173a69908
commit
f6a05e964b
@ -4,9 +4,9 @@
|
|||||||
"cell_type": "markdown",
|
"cell_type": "markdown",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
"# Hugging Face Chat Wrapper\n",
|
"# Hugging Face\n",
|
||||||
"\n",
|
"\n",
|
||||||
"This notebook shows how to get started using Hugging Face LLM's as chat models.\n",
|
"This notebook shows how to get started using `Hugging Face` LLM's as chat models.\n",
|
||||||
"\n",
|
"\n",
|
||||||
"In particular, we will:\n",
|
"In particular, we will:\n",
|
||||||
"1. Utilize the [HuggingFaceTextGenInference](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_text_gen_inference.py), [HuggingFaceEndpoint](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_endpoint.py), or [HuggingFaceHub](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_hub.py) integrations to instantiate an `LLM`.\n",
|
"1. Utilize the [HuggingFaceTextGenInference](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_text_gen_inference.py), [HuggingFaceEndpoint](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_endpoint.py), or [HuggingFaceHub](https://github.com/langchain-ai/langchain/blob/master/libs/langchain/langchain/llms/huggingface_hub.py) integrations to instantiate an `LLM`.\n",
|
||||||
@ -49,7 +49,7 @@
|
|||||||
"cell_type": "markdown",
|
"cell_type": "markdown",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
"#### `HuggingFaceTextGenInference`"
|
"### `HuggingFaceTextGenInference`"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -93,7 +93,7 @@
|
|||||||
"cell_type": "markdown",
|
"cell_type": "markdown",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
"#### `HuggingFaceEndpoint`"
|
"### `HuggingFaceEndpoint`"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -121,7 +121,7 @@
|
|||||||
"cell_type": "markdown",
|
"cell_type": "markdown",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
"#### `HuggingFaceHub`"
|
"### `HuggingFaceHub`"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -291,7 +291,7 @@
|
|||||||
"source": [
|
"source": [
|
||||||
"## 3. Take it for a spin as an agent!\n",
|
"## 3. Take it for a spin as an agent!\n",
|
||||||
"\n",
|
"\n",
|
||||||
"Here we'll test out `Zephyr-7B-beta` as a zero-shot ReAct Agent. The example below is taken from [here](https://python.langchain.com/docs/modules/agents/agent_types/react#using-chat-models).\n",
|
"Here we'll test out `Zephyr-7B-beta` as a zero-shot `ReAct` Agent. The example below is taken from [here](https://python.langchain.com/docs/modules/agents/agent_types/react#using-chat-models).\n",
|
||||||
"\n",
|
"\n",
|
||||||
"> Note: To run this section, you'll need to have a [SerpAPI Token](https://serpapi.com/) saved as an environment variable: `SERPAPI_API_KEY`"
|
"> Note: To run this section, you'll need to have a [SerpAPI Token](https://serpapi.com/) saved as an environment variable: `SERPAPI_API_KEY`"
|
||||||
]
|
]
|
||||||
@ -448,7 +448,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.5"
|
"version": "3.10.12"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
@ -58,31 +58,24 @@ See a [usage example](/docs/integrations/llms/huggingface_textgen_inference).
|
|||||||
from langchain_community.llms import HuggingFaceTextGenInference
|
from langchain_community.llms import HuggingFaceTextGenInference
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Chat models
|
||||||
|
|
||||||
|
### Models from Hugging Face
|
||||||
|
|
||||||
## Document Loaders
|
We can use the `Hugging Face` LLM classes or directly use the `ChatHuggingFace` class.
|
||||||
|
|
||||||
### Hugging Face dataset
|
We need to install several python packages.
|
||||||
|
|
||||||
>[Hugging Face Hub](https://huggingface.co/docs/hub/index) is home to over 75,000
|
|
||||||
> [datasets](https://huggingface.co/docs/hub/index#datasets) in more than 100 languages
|
|
||||||
> that can be used for a broad range of tasks across NLP, Computer Vision, and Audio.
|
|
||||||
> They used for a diverse range of tasks such as translation, automatic speech
|
|
||||||
> recognition, and image classification.
|
|
||||||
|
|
||||||
We need to install `datasets` python package.
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install datasets
|
pip install huggingface_hub
|
||||||
|
pip install transformers
|
||||||
```
|
```
|
||||||
|
See a [usage example](/docs/integrations/chat/huggingface).
|
||||||
See a [usage example](/docs/integrations/document_loaders/hugging_face_dataset).
|
|
||||||
|
|
||||||
```python
|
```python
|
||||||
from langchain_community.document_loaders.hugging_face_dataset import HuggingFaceDatasetLoader
|
from langchain_community.chat_models.huggingface import ChatHuggingFace
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
## Embedding Models
|
## Embedding Models
|
||||||
|
|
||||||
### Hugging Face Hub
|
### Hugging Face Hub
|
||||||
@ -126,6 +119,48 @@ See a [usage example](/docs/integrations/text_embedding/bge_huggingface).
|
|||||||
from langchain_community.embeddings import HuggingFaceBgeEmbeddings
|
from langchain_community.embeddings import HuggingFaceBgeEmbeddings
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Hugging Face Text Embeddings Inference (TEI)
|
||||||
|
|
||||||
|
>[Hugging Face Text Embeddings Inference (TEI)](https://huggingface.co/docs/text-generation-inference/index) is a toolkit for deploying and serving open-source
|
||||||
|
> text embeddings and sequence classification models. `TEI` enables high-performance extraction for the most popular models,
|
||||||
|
>including `FlagEmbedding`, `Ember`, `GTE` and `E5`.
|
||||||
|
|
||||||
|
We need to install `huggingface-hub` python package.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
pip install huggingface-hub
|
||||||
|
```
|
||||||
|
|
||||||
|
See a [usage example](/docs/integrations/text_embedding/text_embeddings_inference).
|
||||||
|
|
||||||
|
```python
|
||||||
|
from langchain_community.embeddings import HuggingFaceHubEmbeddings
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
|
## Document Loaders
|
||||||
|
|
||||||
|
### Hugging Face dataset
|
||||||
|
|
||||||
|
>[Hugging Face Hub](https://huggingface.co/docs/hub/index) is home to over 75,000
|
||||||
|
> [datasets](https://huggingface.co/docs/hub/index#datasets) in more than 100 languages
|
||||||
|
> that can be used for a broad range of tasks across NLP, Computer Vision, and Audio.
|
||||||
|
> They used for a diverse range of tasks such as translation, automatic speech
|
||||||
|
> recognition, and image classification.
|
||||||
|
|
||||||
|
We need to install `datasets` python package.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
pip install datasets
|
||||||
|
```
|
||||||
|
|
||||||
|
See a [usage example](/docs/integrations/document_loaders/hugging_face_dataset).
|
||||||
|
|
||||||
|
```python
|
||||||
|
from langchain_community.document_loaders.hugging_face_dataset import HuggingFaceDatasetLoader
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
## Tools
|
## Tools
|
||||||
|
|
||||||
|
@ -7,7 +7,9 @@
|
|||||||
"source": [
|
"source": [
|
||||||
"# Text Embeddings Inference\n",
|
"# Text Embeddings Inference\n",
|
||||||
"\n",
|
"\n",
|
||||||
"Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5.\n",
|
">[Hugging Face Text Embeddings Inference (TEI)](https://huggingface.co/docs/text-generation-inference/index) is a toolkit for deploying and serving open-source\n",
|
||||||
|
"> text embeddings and sequence classification models. `TEI` enables high-performance extraction for the most popular models,\n",
|
||||||
|
">including `FlagEmbedding`, `Ember`, `GTE` and `E5`.\n",
|
||||||
"\n",
|
"\n",
|
||||||
"To use it within langchain, first install `huggingface-hub`."
|
"To use it within langchain, first install `huggingface-hub`."
|
||||||
]
|
]
|
||||||
@ -21,7 +23,7 @@
|
|||||||
},
|
},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
"source": [
|
"source": [
|
||||||
"%pip install --upgrade --quiet huggingface-hub -q"
|
"%pip install --upgrade huggingface-hub"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -146,9 +148,9 @@
|
|||||||
],
|
],
|
||||||
"metadata": {
|
"metadata": {
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "conda_python3",
|
"display_name": "Python 3 (ipykernel)",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "conda_python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
"language_info": {
|
"language_info": {
|
||||||
"codemirror_mode": {
|
"codemirror_mode": {
|
||||||
@ -160,7 +162,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.13"
|
"version": "3.10.12"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
Loading…
Reference in New Issue
Block a user