mirror of
https://github.com/hwchase17/langchain.git
synced 2025-06-20 05:43:55 +00:00
docs: use init_chat_model (#29623)
This commit is contained in:
parent
bff25b552c
commit
eb9eddae0c
@ -270,7 +270,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs openaiParams={`model=\"gpt-4\"`} />\n"
|
"<ChatModelTabs overrideParams={{openai: {model: \"gpt-4\"}}} />\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
@ -354,7 +354,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" openaiParams={`model=\"gpt-4-0125-preview\", temperature=0`}\n",
|
" overrideParams={{openai: {model: \"gpt-4-0125-preview\", kwargs: \"temperature=0\"}}}\n",
|
||||||
"/>\n"
|
"/>\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
@ -179,7 +179,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" openaiParams={`model=\"gpt-4o\", temperature=0`}\n",
|
" overrideParams={{openai: {model: \"gpt-4o\", kwargs: \"temperature=0\"}}}\n",
|
||||||
"/>\n"
|
"/>\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
@ -167,7 +167,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" fireworksParams={`model=\"accounts/fireworks/models/firefunction-v1\", temperature=0`}\n",
|
" overrideParams={{fireworks: {model: \"accounts/fireworks/models/firefunction-v1\", kwargs: \"temperature=0\"}}}\n",
|
||||||
"/>\n",
|
"/>\n",
|
||||||
"\n",
|
"\n",
|
||||||
"We can use the `bind_tools()` method to handle converting\n",
|
"We can use the `bind_tools()` method to handle converting\n",
|
||||||
|
@ -200,7 +200,12 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" fireworksParams={`model=\"accounts/fireworks/models/firefunction-v1\", temperature=0`}\n",
|
" overrideParams={{\n",
|
||||||
|
" fireworks: {\n",
|
||||||
|
" model: \"accounts/fireworks/models/firefunction-v1\",\n",
|
||||||
|
" kwargs: \"temperature=0\",\n",
|
||||||
|
" }\n",
|
||||||
|
" }}\n",
|
||||||
"/>\n"
|
"/>\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
@ -33,7 +33,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" fireworksParams={`model=\"accounts/fireworks/models/firefunction-v1\", temperature=0`}\n",
|
" overrideParams={{fireworks: {model: \"accounts/fireworks/models/firefunction-v1\", kwargs: \"temperature=0\"}}}\n",
|
||||||
"/>\n"
|
"/>\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
@ -46,7 +46,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs\n",
|
"<ChatModelTabs\n",
|
||||||
" customVarName=\"llm\"\n",
|
" customVarName=\"llm\"\n",
|
||||||
" fireworksParams={`model=\"accounts/fireworks/models/firefunction-v1\", temperature=0`}\n",
|
" overrideParams={{fireworks: {model: \"accounts/fireworks/models/firefunction-v1\", kwargs: \"temperature=0\"}}}\n",
|
||||||
"/>\n"
|
"/>\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
@ -91,7 +91,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs openaiParams={`model=\"gpt-4\"`} />\n",
|
"<ChatModelTabs overrideParams={{openai: {model: \"gpt-4\"}}} />\n",
|
||||||
"\n",
|
"\n",
|
||||||
"To illustrate the idea, we'll use `phi3` via Ollama, which does **NOT** have native support for tool calling. If you'd like to use `Ollama` as well follow [these instructions](/docs/integrations/chat/ollama/)."
|
"To illustrate the idea, we'll use `phi3` via Ollama, which does **NOT** have native support for tool calling. If you'd like to use `Ollama` as well follow [these instructions](/docs/integrations/chat/ollama/)."
|
||||||
]
|
]
|
||||||
|
@ -17,7 +17,7 @@ If you'd like to contribute an integration, see [Contributing integrations](/doc
|
|||||||
|
|
||||||
import ChatModelTabs from "@theme/ChatModelTabs";
|
import ChatModelTabs from "@theme/ChatModelTabs";
|
||||||
|
|
||||||
<ChatModelTabs openaiParams={`model="gpt-4o-mini"`} />
|
<ChatModelTabs overrideParams={{openai: {model: "gpt-4o-mini"}}} />
|
||||||
|
|
||||||
```python
|
```python
|
||||||
model.invoke("Hello, world!")
|
model.invoke("Hello, world!")
|
||||||
|
@ -1,34 +0,0 @@
|
|||||||
# FalkorDB
|
|
||||||
|
|
||||||
>[FalkorDB](https://www.falkordb.com/) is a creator of the [FalkorDB](https://docs.falkordb.com/),
|
|
||||||
> a low-latency Graph Database that delivers knowledge to GenAI.
|
|
||||||
|
|
||||||
|
|
||||||
## Installation and Setup
|
|
||||||
|
|
||||||
See [installation instructions here](/docs/integrations/graphs/falkordb/).
|
|
||||||
|
|
||||||
|
|
||||||
## Graphs
|
|
||||||
|
|
||||||
See a [usage example](/docs/integrations/graphs/falkordb).
|
|
||||||
|
|
||||||
```python
|
|
||||||
from langchain_community.graphs import FalkorDBGraph
|
|
||||||
```
|
|
||||||
|
|
||||||
## Chains
|
|
||||||
|
|
||||||
See a [usage example](/docs/integrations/graphs/falkordb).
|
|
||||||
|
|
||||||
```python
|
|
||||||
from langchain_community.chains.graph_qa.falkordb import FalkorDBQAChain
|
|
||||||
```
|
|
||||||
|
|
||||||
## Memory
|
|
||||||
|
|
||||||
See a [usage example](/docs/integrations/memory/falkordb_chat_message_history).
|
|
||||||
|
|
||||||
```python
|
|
||||||
from langchain_falkordb import FalkorDBChatMessageHistory
|
|
||||||
```
|
|
@ -215,7 +215,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs openaiParams={`model=\"gpt-4\"`} />\n"
|
"<ChatModelTabs overrideParams={{openai: {model: \"gpt-4\"}}} />\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
@ -108,7 +108,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs openaiParams={`model=\"gpt-4o-mini\"`} />\n"
|
"<ChatModelTabs overrideParams={{openai: {model: \"gpt-4o-mini\"}}} />\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -935,7 +935,7 @@
|
|||||||
],
|
],
|
||||||
"metadata": {
|
"metadata": {
|
||||||
"kernelspec": {
|
"kernelspec": {
|
||||||
"display_name": "Python 3 (ipykernel)",
|
"display_name": ".venv",
|
||||||
"language": "python",
|
"language": "python",
|
||||||
"name": "python3"
|
"name": "python3"
|
||||||
},
|
},
|
||||||
@ -949,7 +949,7 @@
|
|||||||
"name": "python",
|
"name": "python",
|
||||||
"nbconvert_exporter": "python",
|
"nbconvert_exporter": "python",
|
||||||
"pygments_lexer": "ipython3",
|
"pygments_lexer": "ipython3",
|
||||||
"version": "3.10.4"
|
"version": "3.11.4"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nbformat": 4,
|
"nbformat": 4,
|
||||||
|
@ -91,7 +91,7 @@
|
|||||||
"\n",
|
"\n",
|
||||||
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
"import ChatModelTabs from \"@theme/ChatModelTabs\";\n",
|
||||||
"\n",
|
"\n",
|
||||||
"<ChatModelTabs openaiParams={`model=\"gpt-4o-mini\"`} />\n"
|
"<ChatModelTabs overrideParams={{openai: {model: \"gpt-4o-mini\"}}} />\n"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
@ -194,7 +194,7 @@
|
|||||||
"id": "4c0766af-a3b3-4293-b253-3a10f365ab5d",
|
"id": "4c0766af-a3b3-4293-b253-3a10f365ab5d",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"source": [
|
"source": [
|
||||||
":::hint\n",
|
":::tip\n",
|
||||||
"\n",
|
"\n",
|
||||||
"This also supports streaming LLM content token by token if using langgraph >= 0.2.28.\n",
|
"This also supports streaming LLM content token by token if using langgraph >= 0.2.28.\n",
|
||||||
":::"
|
":::"
|
||||||
|
@ -91,29 +91,7 @@ export const CustomDropdown = ({ selectedOption, options, onSelect, modelType })
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* @typedef {Object} ChatModelTabsProps - Component props.
|
* @typedef {Object} ChatModelTabsProps - Component props.
|
||||||
* @property {string} [openaiParams] - Parameters for OpenAI chat model. Defaults to `model="gpt-3.5-turbo-0125"`
|
* @property {Object} [overrideParams] - An object for overriding the default parameters for each chat model, e.g. `{ openai: { model: "gpt-4o-mini" } }`
|
||||||
* @property {string} [anthropicParams] - Parameters for Anthropic chat model. Defaults to `model="claude-3-sonnet-20240229"`
|
|
||||||
* @property {string} [cohereParams] - Parameters for Cohere chat model. Defaults to `model="command-r-plus"`
|
|
||||||
* @property {string} [fireworksParams] - Parameters for Fireworks chat model. Defaults to `model="accounts/fireworks/models/mixtral-8x7b-instruct"`
|
|
||||||
* @property {string} [groqParams] - Parameters for Groq chat model. Defaults to `model="llama3-8b-8192"`
|
|
||||||
* @property {string} [mistralParams] - Parameters for Mistral chat model. Defaults to `model="mistral-large-latest"`
|
|
||||||
* @property {string} [googleParams] - Parameters for Google chat model. Defaults to `model="gemini-pro"`
|
|
||||||
* @property {string} [togetherParams] - Parameters for Together chat model. Defaults to `model="mistralai/Mixtral-8x7B-Instruct-v0.1"`
|
|
||||||
* @property {string} [nvidiaParams] - Parameters for Nvidia NIM model. Defaults to `model="meta/llama3-70b-instruct"`
|
|
||||||
* @property {string} [databricksParams] - Parameters for Databricks model. Defaults to `endpoint="databricks-meta-llama-3-1-70b-instruct"`
|
|
||||||
* @property {string} [awsBedrockParams] - Parameters for AWS Bedrock chat model.
|
|
||||||
* @property {boolean} [hideOpenai] - Whether or not to hide OpenAI chat model.
|
|
||||||
* @property {boolean} [hideAnthropic] - Whether or not to hide Anthropic chat model.
|
|
||||||
* @property {boolean} [hideCohere] - Whether or not to hide Cohere chat model.
|
|
||||||
* @property {boolean} [hideFireworks] - Whether or not to hide Fireworks chat model.
|
|
||||||
* @property {boolean} [hideGroq] - Whether or not to hide Groq chat model.
|
|
||||||
* @property {boolean} [hideMistral] - Whether or not to hide Mistral chat model.
|
|
||||||
* @property {boolean} [hideGoogle] - Whether or not to hide Google VertexAI chat model.
|
|
||||||
* @property {boolean} [hideTogether] - Whether or not to hide Together chat model.
|
|
||||||
* @property {boolean} [hideAzure] - Whether or not to hide Microsoft Azure OpenAI chat model.
|
|
||||||
* @property {boolean} [hideNvidia] - Whether or not to hide NVIDIA NIM model.
|
|
||||||
* @property {boolean} [hideAWS] - Whether or not to hide AWS models.
|
|
||||||
* @property {boolean} [hideDatabricks] - Whether or not to hide Databricks models.
|
|
||||||
* @property {string} [customVarName] - Custom variable name for the model. Defaults to `model`.
|
* @property {string} [customVarName] - Custom variable name for the model. Defaults to `model`.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
@ -121,198 +99,151 @@ export const CustomDropdown = ({ selectedOption, options, onSelect, modelType })
|
|||||||
* @param {ChatModelTabsProps} props - Component props.
|
* @param {ChatModelTabsProps} props - Component props.
|
||||||
*/
|
*/
|
||||||
export default function ChatModelTabs(props) {
|
export default function ChatModelTabs(props) {
|
||||||
const [selectedModel, setSelectedModel] = useState("Groq");
|
const [selectedModel, setSelectedModel] = useState("groq");
|
||||||
const {
|
const {
|
||||||
openaiParams,
|
overrideParams,
|
||||||
anthropicParams,
|
|
||||||
cohereParams,
|
|
||||||
fireworksParams,
|
|
||||||
groqParams,
|
|
||||||
mistralParams,
|
|
||||||
googleParams,
|
|
||||||
togetherParams,
|
|
||||||
azureParams,
|
|
||||||
nvidiaParams,
|
|
||||||
awsBedrockParams,
|
|
||||||
databricksParams,
|
|
||||||
hideOpenai,
|
|
||||||
hideAnthropic,
|
|
||||||
hideCohere,
|
|
||||||
hideFireworks,
|
|
||||||
hideGroq,
|
|
||||||
hideMistral,
|
|
||||||
hideGoogle,
|
|
||||||
hideTogether,
|
|
||||||
hideAzure,
|
|
||||||
hideNvidia,
|
|
||||||
hideAWS,
|
|
||||||
hideDatabricks,
|
|
||||||
customVarName,
|
customVarName,
|
||||||
} = props;
|
} = props;
|
||||||
|
|
||||||
const openAIParamsOrDefault = openaiParams ?? `model="gpt-4o-mini"`;
|
|
||||||
const anthropicParamsOrDefault =
|
|
||||||
anthropicParams ?? `model="claude-3-5-sonnet-20240620"`;
|
|
||||||
const cohereParamsOrDefault = cohereParams ?? `model="command-r-plus"`;
|
|
||||||
const fireworksParamsOrDefault =
|
|
||||||
fireworksParams ??
|
|
||||||
`model="accounts/fireworks/models/llama-v3p1-70b-instruct"`;
|
|
||||||
const groqParamsOrDefault = groqParams ?? `model="llama3-8b-8192"`;
|
|
||||||
const mistralParamsOrDefault =
|
|
||||||
mistralParams ?? `model="mistral-large-latest"`;
|
|
||||||
const googleParamsOrDefault = googleParams ?? `model="gemini-1.5-flash"`;
|
|
||||||
const togetherParamsOrDefault =
|
|
||||||
togetherParams ??
|
|
||||||
`\n base_url="https://api.together.xyz/v1",\n api_key=os.environ["TOGETHER_API_KEY"],\n model="mistralai/Mixtral-8x7B-Instruct-v0.1",\n`;
|
|
||||||
const azureParamsOrDefault =
|
|
||||||
azureParams ??
|
|
||||||
`\n azure_endpoint=os.environ["AZURE_OPENAI_ENDPOINT"],\n azure_deployment=os.environ["AZURE_OPENAI_DEPLOYMENT_NAME"],\n openai_api_version=os.environ["AZURE_OPENAI_API_VERSION"],\n`;
|
|
||||||
const nvidiaParamsOrDefault = nvidiaParams ?? `model="meta/llama3-70b-instruct"`
|
|
||||||
const awsBedrockParamsOrDefault = awsBedrockParams ?? `model="anthropic.claude-3-5-sonnet-20240620-v1:0",\n beta_use_converse_api=True`;
|
|
||||||
const databricksParamsOrDefault = databricksParams ?? `endpoint="databricks-meta-llama-3-1-70b-instruct"`
|
|
||||||
|
|
||||||
const llmVarName = customVarName ?? "model";
|
const llmVarName = customVarName ?? "model";
|
||||||
|
|
||||||
const tabItems = [
|
const tabItems = [
|
||||||
{
|
{
|
||||||
value: "Groq",
|
value: "groq",
|
||||||
label: "Groq",
|
label: "Groq",
|
||||||
text: `from langchain_groq import ChatGroq\n\n${llmVarName} = ChatGroq(${groqParamsOrDefault})`,
|
model: "llama3-8b-8192",
|
||||||
apiKeyName: "GROQ_API_KEY",
|
apiKeyName: "GROQ_API_KEY",
|
||||||
packageName: "langchain-groq",
|
packageName: "langchain-groq",
|
||||||
shouldHide: hideGroq,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "OpenAI",
|
value: "openai",
|
||||||
label: "OpenAI",
|
label: "OpenAI",
|
||||||
text: `from langchain_openai import ChatOpenAI\n\n${llmVarName} = ChatOpenAI(${openAIParamsOrDefault})`,
|
model: "gpt-4o-mini",
|
||||||
apiKeyName: "OPENAI_API_KEY",
|
apiKeyName: "OPENAI_API_KEY",
|
||||||
packageName: "langchain-openai",
|
packageName: "langchain-openai",
|
||||||
shouldHide: hideOpenai,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "Anthropic",
|
value: "anthropic",
|
||||||
label: "Anthropic",
|
label: "Anthropic",
|
||||||
text: `from langchain_anthropic import ChatAnthropic\n\n${llmVarName} = ChatAnthropic(${anthropicParamsOrDefault})`,
|
model: "claude-3-5-sonnet-latest",
|
||||||
apiKeyName: "ANTHROPIC_API_KEY",
|
apiKeyName: "ANTHROPIC_API_KEY",
|
||||||
packageName: "langchain-anthropic",
|
packageName: "langchain-anthropic",
|
||||||
shouldHide: hideAnthropic,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "Azure",
|
value: "azure",
|
||||||
label: "Azure",
|
label: "Azure",
|
||||||
text: `from langchain_openai import AzureChatOpenAI\n\n${llmVarName} = AzureChatOpenAI(${azureParamsOrDefault})`,
|
text: `from langchain_openai import AzureChatOpenAI
|
||||||
|
|
||||||
|
${llmVarName} = AzureChatOpenAI(
|
||||||
|
azure_endpoint=os.environ["AZURE_OPENAI_ENDPOINT"],
|
||||||
|
azure_deployment=os.environ["AZURE_OPENAI_DEPLOYMENT_NAME"],
|
||||||
|
openai_api_version=os.environ["AZURE_OPENAI_API_VERSION"],
|
||||||
|
)`,
|
||||||
apiKeyName: "AZURE_OPENAI_API_KEY",
|
apiKeyName: "AZURE_OPENAI_API_KEY",
|
||||||
packageName: "langchain-openai",
|
packageName: "langchain-openai",
|
||||||
shouldHide: hideAzure,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "Google",
|
value: "google_vertexai",
|
||||||
label: "Google",
|
label: "Google Vertex",
|
||||||
text: `from langchain_google_vertexai import ChatVertexAI\n\n${llmVarName} = ChatVertexAI(${googleParamsOrDefault})`,
|
model: "gemini-2.0-flash",
|
||||||
apiKeyText: "# Ensure your VertexAI credentials are configured",
|
apiKeyText: "# Ensure your VertexAI credentials are configured",
|
||||||
packageName: "langchain-google-vertexai",
|
packageName: "langchain-google-vertexai",
|
||||||
shouldHide: hideGoogle,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "AWS",
|
value: "aws",
|
||||||
label: "AWS",
|
label: "AWS",
|
||||||
text: `from langchain_aws import ChatBedrock\n\n${llmVarName} = ChatBedrock(${awsBedrockParamsOrDefault})`,
|
model: "anthropic.claude-3-5-sonnet-20240620-v1:0",
|
||||||
|
kwargs: "beta_use_converse_api=True",
|
||||||
apiKeyText: "# Ensure your AWS credentials are configured",
|
apiKeyText: "# Ensure your AWS credentials are configured",
|
||||||
packageName: "langchain-aws",
|
packageName: "langchain-aws",
|
||||||
shouldHide: hideAWS,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "Cohere",
|
value: "cohere",
|
||||||
label: "Cohere",
|
label: "Cohere",
|
||||||
text: `from langchain_cohere import ChatCohere\n\n${llmVarName} = ChatCohere(${cohereParamsOrDefault})`,
|
model: "command-r-plus",
|
||||||
apiKeyName: "COHERE_API_KEY",
|
apiKeyName: "COHERE_API_KEY",
|
||||||
packageName: "langchain-cohere",
|
packageName: "langchain-cohere",
|
||||||
shouldHide: hideCohere,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "NVIDIA",
|
value: "nvidia",
|
||||||
label: "NVIDIA",
|
label: "NVIDIA",
|
||||||
text: `from langchain_nvidia_ai_endpoints import ChatNVIDIA\n\n${llmVarName} = ChatNVIDIA(${nvidiaParamsOrDefault})`,
|
model: "meta/llama3-70b-instruct",
|
||||||
apiKeyName: "NVIDIA_API_KEY",
|
apiKeyName: "NVIDIA_API_KEY",
|
||||||
packageName: "langchain-nvidia-ai-endpoints",
|
packageName: "langchain-nvidia-ai-endpoints",
|
||||||
shouldHide: hideNvidia,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "FireworksAI",
|
value: "fireworks",
|
||||||
label: "Fireworks AI",
|
label: "Fireworks AI",
|
||||||
text: `from langchain_fireworks import ChatFireworks\n\n${llmVarName} = ChatFireworks(${fireworksParamsOrDefault})`,
|
model: "accounts/fireworks/models/llama-v3p1-70b-instruct",
|
||||||
apiKeyName: "FIREWORKS_API_KEY",
|
apiKeyName: "FIREWORKS_API_KEY",
|
||||||
packageName: "langchain-fireworks",
|
packageName: "langchain-fireworks",
|
||||||
shouldHide: hideFireworks,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "MistralAI",
|
value: "mistralai",
|
||||||
label: "Mistral AI",
|
label: "Mistral AI",
|
||||||
text: `from langchain_mistralai import ChatMistralAI\n\n${llmVarName} = ChatMistralAI(${mistralParamsOrDefault})`,
|
model: "mistral-large-latest",
|
||||||
apiKeyName: "MISTRAL_API_KEY",
|
apiKeyName: "MISTRAL_API_KEY",
|
||||||
packageName: "langchain-mistralai",
|
packageName: "langchain-mistralai",
|
||||||
shouldHide: hideMistral,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "TogetherAI",
|
value: "together",
|
||||||
label: "Together AI",
|
label: "Together AI",
|
||||||
text: `from langchain_openai import ChatOpenAI\n\n${llmVarName} = ChatOpenAI(${togetherParamsOrDefault})`,
|
model: "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||||
apiKeyName: "TOGETHER_API_KEY",
|
apiKeyName: "TOGETHER_API_KEY",
|
||||||
packageName: "langchain-openai",
|
packageName: "langchain-together",
|
||||||
shouldHide: hideTogether,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
value: "Databricks",
|
value: "databricks",
|
||||||
label: "Databricks",
|
label: "Databricks",
|
||||||
text: `from databricks_langchain import ChatDatabricks\n\nos.environ["DATABRICKS_HOST"] = "https://example.staging.cloud.databricks.com/serving-endpoints"\n\n${llmVarName} = ChatDatabricks(${databricksParamsOrDefault})`,
|
text: `from databricks_langchain import ChatDatabricks\n\nos.environ["DATABRICKS_HOST"] = "https://example.staging.cloud.databricks.com/serving-endpoints"\n\n${llmVarName} = ChatDatabricks(endpoint="databricks-meta-llama-3-1-70b-instruct")`,
|
||||||
apiKeyName: "DATABRICKS_TOKEN",
|
apiKeyName: "DATABRICKS_TOKEN",
|
||||||
packageName: "databricks-langchain",
|
packageName: "databricks-langchain",
|
||||||
shouldHide: hideDatabricks,
|
|
||||||
},
|
},
|
||||||
];
|
].map((item) => ({
|
||||||
|
...item,
|
||||||
|
...overrideParams?.[item.value],
|
||||||
|
}));
|
||||||
|
|
||||||
const modelOptions = tabItems
|
const modelOptions = tabItems
|
||||||
.filter((item) => !item.shouldHide)
|
|
||||||
.map((item) => ({
|
.map((item) => ({
|
||||||
value: item.value,
|
value: item.value,
|
||||||
label: item.label,
|
label: item.label,
|
||||||
text: item.text,
|
|
||||||
apiKeyName: item.apiKeyName,
|
|
||||||
apiKeyText: item.apiKeyText,
|
|
||||||
packageName: item.packageName,
|
|
||||||
}));
|
}));
|
||||||
|
|
||||||
const selectedOption = modelOptions.find(
|
const selectedTabItem = tabItems.find(
|
||||||
(option) => option.value === selectedModel
|
(option) => option.value === selectedModel
|
||||||
);
|
);
|
||||||
|
|
||||||
let apiKeyText = "";
|
let apiKeyText = "";
|
||||||
if (selectedOption.apiKeyName) {
|
if (selectedTabItem.apiKeyName) {
|
||||||
apiKeyText = `import getpass
|
apiKeyText = `import getpass
|
||||||
import os
|
import os
|
||||||
|
|
||||||
if not os.environ.get("${selectedOption.apiKeyName}"):
|
if not os.environ.get("${selectedTabItem.apiKeyName}"):
|
||||||
os.environ["${selectedOption.apiKeyName}"] = getpass.getpass("Enter API key for ${selectedOption.label}: ")`;
|
os.environ["${selectedTabItem.apiKeyName}"] = getpass.getpass("Enter API key for ${selectedTabItem.label}: ")`;
|
||||||
} else if (selectedOption.apiKeyText) {
|
} else if (selectedTabItem.apiKeyText) {
|
||||||
apiKeyText = selectedOption.apiKeyText;
|
apiKeyText = selectedTabItem.apiKeyText;
|
||||||
}
|
}
|
||||||
|
|
||||||
return (
|
const initModelText = selectedTabItem?.text || `from langchain.chat_models import init_chat_model
|
||||||
<div>
|
|
||||||
<CustomDropdown
|
|
||||||
selectedOption={selectedOption}
|
|
||||||
options={modelOptions}
|
|
||||||
onSelect={setSelectedModel}
|
|
||||||
modelType="chat"
|
|
||||||
/>
|
|
||||||
|
|
||||||
<CodeBlock language="bash">
|
${llmVarName} = init_chat_model("${selectedTabItem.model}", *, model_provider="${selectedTabItem.value}"${selectedTabItem?.kwargs ? `, ${selectedTabItem.kwargs}` : ""})`;
|
||||||
{`pip install -qU ${selectedOption.packageName}`}
|
|
||||||
</CodeBlock>
|
return (
|
||||||
<CodeBlock language="python">
|
<div>
|
||||||
{apiKeyText ? apiKeyText + "\n\n" + selectedOption.text : selectedOption.text}
|
<CustomDropdown
|
||||||
</CodeBlock>
|
selectedOption={selectedTabItem}
|
||||||
</div>
|
options={modelOptions}
|
||||||
);
|
onSelect={setSelectedModel}
|
||||||
|
modelType="chat"
|
||||||
|
/>
|
||||||
|
|
||||||
|
<CodeBlock language="bash">
|
||||||
|
{`pip install -qU langchain ${selectedTabItem.packageName}`}
|
||||||
|
</CodeBlock>
|
||||||
|
<CodeBlock language="python">
|
||||||
|
{apiKeyText ? apiKeyText + "\n\n" + initModelText : initModelText}
|
||||||
|
</CodeBlock>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
}
|
}
|
@ -117,6 +117,8 @@ def init_chat_model(
|
|||||||
- 'groq' -> langchain-groq
|
- 'groq' -> langchain-groq
|
||||||
- 'ollama' -> langchain-ollama
|
- 'ollama' -> langchain-ollama
|
||||||
- 'google_anthropic_vertex' -> langchain-google-vertexai
|
- 'google_anthropic_vertex' -> langchain-google-vertexai
|
||||||
|
- 'deepseek' -> langchain-deepseek
|
||||||
|
- 'nvidia' -> langchain-nvidia-ai-endpoints
|
||||||
|
|
||||||
Will attempt to infer model_provider from model if not specified. The
|
Will attempt to infer model_provider from model if not specified. The
|
||||||
following providers will be inferred based on these model prefixes:
|
following providers will be inferred based on these model prefixes:
|
||||||
@ -421,6 +423,11 @@ def _init_chat_model_helper(
|
|||||||
from langchain_deepseek import ChatDeepSeek
|
from langchain_deepseek import ChatDeepSeek
|
||||||
|
|
||||||
return ChatDeepSeek(model=model, **kwargs)
|
return ChatDeepSeek(model=model, **kwargs)
|
||||||
|
elif model_provider == "nvidia":
|
||||||
|
_check_pkg("langchain_nvidia_ai_endpoints")
|
||||||
|
from langchain_nvidia_ai_endpoints import ChatNVIDIA
|
||||||
|
|
||||||
|
return ChatNVIDIA(model=model, **kwargs)
|
||||||
else:
|
else:
|
||||||
supported = ", ".join(_SUPPORTED_PROVIDERS)
|
supported = ", ".join(_SUPPORTED_PROVIDERS)
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
|
Loading…
Reference in New Issue
Block a user