mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-06-29 08:47:19 +00:00
Merge branch 'feature/upgrade-llamaindex' into feature/tensorrt-support
This commit is contained in:
commit
3aaf4d682b
@ -62,6 +62,23 @@ Where `<extra>` can be any of the following:
|
|||||||
There are just some examples of recommended setups. You can mix and match the different options to fit your needs.
|
There are just some examples of recommended setups. You can mix and match the different options to fit your needs.
|
||||||
You'll find more information in the Manual section of the documentation.
|
You'll find more information in the Manual section of the documentation.
|
||||||
|
|
||||||
|
> **Important for Windows**: In the examples below or how to run PrivateGPT with `make run`, `PGPT_PROFILES` env var is being set inline following Unix command line syntax (works on MacOS and Linux).
|
||||||
|
If you are using Windows, you'll need to set the env var in a different way, for example:
|
||||||
|
|
||||||
|
```powershell
|
||||||
|
# Powershell
|
||||||
|
$env:PGPT_PROFILES="ollama"
|
||||||
|
make run
|
||||||
|
```
|
||||||
|
|
||||||
|
or
|
||||||
|
|
||||||
|
```cmd
|
||||||
|
# CMD
|
||||||
|
set PGPT_PROFILES=ollama
|
||||||
|
make run
|
||||||
|
```
|
||||||
|
|
||||||
### Local, Ollama-powered setup
|
### Local, Ollama-powered setup
|
||||||
|
|
||||||
The easiest way to run PrivateGPT fully locally is to depend on Ollama for the LLM. Ollama provides a local LLM that is easy to install and use.
|
The easiest way to run PrivateGPT fully locally is to depend on Ollama for the LLM. Ollama provides a local LLM that is easy to install and use.
|
||||||
|
@ -87,7 +87,7 @@ class LLMComponent:
|
|||||||
from llama_index.llms.openai_like import OpenAILike # type: ignore
|
from llama_index.llms.openai_like import OpenAILike # type: ignore
|
||||||
except ImportError as e:
|
except ImportError as e:
|
||||||
raise ImportError(
|
raise ImportError(
|
||||||
"OpenAILike dependencies not found, install with `poetry install --extras llms-openailike`"
|
"OpenAILike dependencies not found, install with `poetry install --extras llms-openai-like`"
|
||||||
) from e
|
) from e
|
||||||
|
|
||||||
openai_settings = settings.openai
|
openai_settings = settings.openai
|
||||||
|
@ -19,19 +19,19 @@ os.makedirs(models_path, exist_ok=True)
|
|||||||
|
|
||||||
# Download Embedding model
|
# Download Embedding model
|
||||||
embedding_path = models_path / "embedding"
|
embedding_path = models_path / "embedding"
|
||||||
print(f"Downloading embedding {settings().local.embedding_hf_model_name}")
|
print(f"Downloading embedding {settings().huggingface.embedding_hf_model_name}")
|
||||||
snapshot_download(
|
snapshot_download(
|
||||||
repo_id=settings().local.embedding_hf_model_name,
|
repo_id=settings().huggingface.embedding_hf_model_name,
|
||||||
cache_dir=models_cache_path,
|
cache_dir=models_cache_path,
|
||||||
local_dir=embedding_path,
|
local_dir=embedding_path,
|
||||||
)
|
)
|
||||||
print("Embedding model downloaded!")
|
print("Embedding model downloaded!")
|
||||||
|
|
||||||
# Download LLM and create a symlink to the model file
|
# Download LLM and create a symlink to the model file
|
||||||
print(f"Downloading LLM {settings().local.llm_hf_model_file}")
|
print(f"Downloading LLM {settings().llamacpp.llm_hf_model_file}")
|
||||||
hf_hub_download(
|
hf_hub_download(
|
||||||
repo_id=settings().local.llm_hf_repo_id,
|
repo_id=settings().llamacpp.llm_hf_repo_id,
|
||||||
filename=settings().local.llm_hf_model_file,
|
filename=settings().llamacpp.llm_hf_model_file,
|
||||||
cache_dir=models_cache_path,
|
cache_dir=models_cache_path,
|
||||||
local_dir=models_path,
|
local_dir=models_path,
|
||||||
resume_download=resume_download,
|
resume_download=resume_download,
|
||||||
|
Loading…
Reference in New Issue
Block a user