mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-06-19 12:12:35 +00:00
FEAT: Added fixed UI mode name
This commit is contained in:
parent
a38675b844
commit
2d5865e773
4284
poetry.lock
generated
4284
poetry.lock
generated
File diff suppressed because it is too large
Load Diff
@ -80,7 +80,6 @@ class EmbeddingComponent:
|
|||||||
self.embedding_model = FireworksEmbedding(
|
self.embedding_model = FireworksEmbedding(
|
||||||
api_key=api_key,
|
api_key=api_key,
|
||||||
)
|
)
|
||||||
|
|
||||||
case "ollama":
|
case "ollama":
|
||||||
try:
|
try:
|
||||||
from llama_index.embeddings.ollama import ( # type: ignore
|
from llama_index.embeddings.ollama import ( # type: ignore
|
||||||
|
@ -115,7 +115,7 @@ class LLMSettings(BaseModel):
|
|||||||
"mock",
|
"mock",
|
||||||
"ollama",
|
"ollama",
|
||||||
"gemini",
|
"gemini",
|
||||||
"fireworks"
|
"fireworks",
|
||||||
]
|
]
|
||||||
max_new_tokens: int = Field(
|
max_new_tokens: int = Field(
|
||||||
256,
|
256,
|
||||||
@ -198,7 +198,7 @@ class HuggingFaceSettings(BaseModel):
|
|||||||
|
|
||||||
class EmbeddingSettings(BaseModel):
|
class EmbeddingSettings(BaseModel):
|
||||||
mode: Literal[
|
mode: Literal[
|
||||||
"huggingface", "openai", "azopenai", "sagemaker", "ollama", "mock", "gemini"
|
"huggingface", "openai", "azopenai", "sagemaker", "ollama", "mock", "gemini","fireworks"
|
||||||
]
|
]
|
||||||
ingest_mode: Literal["simple", "batch", "parallel", "pipeline"] = Field(
|
ingest_mode: Literal["simple", "batch", "parallel", "pipeline"] = Field(
|
||||||
"simple",
|
"simple",
|
||||||
|
@ -518,6 +518,7 @@ class PrivateGptUi:
|
|||||||
model_mapping = {
|
model_mapping = {
|
||||||
"llamacpp": config_settings.llamacpp.llm_hf_model_file,
|
"llamacpp": config_settings.llamacpp.llm_hf_model_file,
|
||||||
"openai": config_settings.openai.model,
|
"openai": config_settings.openai.model,
|
||||||
|
"fireworks": config_settings.fireworks.model,
|
||||||
"openailike": config_settings.openai.model,
|
"openailike": config_settings.openai.model,
|
||||||
"azopenai": config_settings.azopenai.llm_model,
|
"azopenai": config_settings.azopenai.llm_model,
|
||||||
"sagemaker": config_settings.sagemaker.llm_endpoint_name,
|
"sagemaker": config_settings.sagemaker.llm_endpoint_name,
|
||||||
|
@ -37,9 +37,9 @@ llama-index-vector-stores-postgres = {version ="^0.1.11", optional = true}
|
|||||||
llama-index-vector-stores-clickhouse = {version ="^0.1.3", optional = true}
|
llama-index-vector-stores-clickhouse = {version ="^0.1.3", optional = true}
|
||||||
llama-index-storage-docstore-postgres = {version ="^0.1.3", optional = true}
|
llama-index-storage-docstore-postgres = {version ="^0.1.3", optional = true}
|
||||||
llama-index-storage-index-store-postgres = {version ="^0.1.4", optional = true}
|
llama-index-storage-index-store-postgres = {version ="^0.1.4", optional = true}
|
||||||
# firworks
|
|
||||||
llama-index-llms-fireworks = {version = "^0.1.5", optional = true}
|
llama-index-llms-fireworks = {version = "^0.1.5", optional = true}
|
||||||
llama-index-embeddings-fireworks = {version = "^0.1.2", optional = true}
|
llama-index-embeddings-fireworks = {version = "^0.1.2", optional = true}
|
||||||
|
|
||||||
# Postgres
|
# Postgres
|
||||||
psycopg2-binary = {version ="^2.9.9", optional = true}
|
psycopg2-binary = {version ="^2.9.9", optional = true}
|
||||||
asyncpg = {version="^0.29.0", optional = true}
|
asyncpg = {version="^0.29.0", optional = true}
|
||||||
|
@ -10,4 +10,4 @@ embedding:
|
|||||||
fireworks:
|
fireworks:
|
||||||
api_key: ${FIREWORKS_API_KEY:}
|
api_key: ${FIREWORKS_API_KEY:}
|
||||||
model: "accounts/fireworks/models/llama-v3p1-70b-instruct"
|
model: "accounts/fireworks/models/llama-v3p1-70b-instruct"
|
||||||
#poetry install --extras "ui llms-fireworks embeddings-fireworks vector-stores-qdrant"
|
#poetry install --extras "ui llms-fireworks embeddings-fireworks vector-stores-qdrant embeddings-openai"
|
||||||
|
Loading…
Reference in New Issue
Block a user