privateGPT/private_gpt
Javier Martinez e54a8fe043
fix: prevent to ingest local files (by default) (#2010)
* feat: prevent to local ingestion (by default) and add white-list

* docs: add local ingestion warning

* docs: add missing comment

* fix: update exception error

* fix: black
2024-07-31 14:33:46 +02:00
..
components added llama3 prompt (#1962) 2024-07-29 17:28:00 +02:00
open_ai feat: Upgrade to LlamaIndex to 0.10 (#1663) 2024-03-06 17:51:30 +01:00
server feat(RAG): Introduce SentenceTransformer Reranker (#1810) 2024-04-02 10:29:51 +02:00
settings fix: prevent to ingest local files (by default) (#2010) 2024-07-31 14:33:46 +02:00
ui fix: light mode (#2025) 2024-07-31 12:59:31 +02:00
utils feat(llm): autopull ollama models (#2019) 2024-07-29 13:25:42 +02:00
__init__.py feat(local): tiktoken cache within repo for offline (#1467) 2024-03-11 22:55:13 +01:00
__main__.py fix: Remove global state (#1216) 2023-11-12 22:20:36 +01:00
constants.py Next version of PrivateGPT (#1077) 2023-10-19 16:04:35 +02:00
di.py fix: Remove global state (#1216) 2023-11-12 22:20:36 +01:00
launcher.py feat(llm): adds serveral settings for llamacpp and ollama (#1703) 2024-03-11 22:51:05 +01:00
main.py feat: Upgrade to LlamaIndex to 0.10 (#1663) 2024-03-06 17:51:30 +01:00
paths.py fix: Remove global state (#1216) 2023-11-12 22:20:36 +01:00