mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-06-28 00:08:39 +00:00
Format fixes
This commit is contained in:
parent
8c390812ff
commit
34d48d7b4d
@ -1,6 +1,5 @@
|
|||||||
import logging
|
import logging
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Dict, Type
|
|
||||||
|
|
||||||
from llama_index.core.readers import StringIterableReader
|
from llama_index.core.readers import StringIterableReader
|
||||||
from llama_index.core.readers.base import BaseReader
|
from llama_index.core.readers.base import BaseReader
|
||||||
@ -11,7 +10,7 @@ logger = logging.getLogger(__name__)
|
|||||||
|
|
||||||
|
|
||||||
# Inspired by the `llama_index.core.readers.file.base` module
|
# Inspired by the `llama_index.core.readers.file.base` module
|
||||||
def _try_loading_included_file_formats() -> Dict[str, Type[BaseReader]]:
|
def _try_loading_included_file_formats() -> dict[str, type[BaseReader]]:
|
||||||
try:
|
try:
|
||||||
from llama_index.readers.file.docs import DocxReader, HWPReader, PDFReader
|
from llama_index.readers.file.docs import DocxReader, HWPReader, PDFReader
|
||||||
from llama_index.readers.file.epub import EpubReader
|
from llama_index.readers.file.epub import EpubReader
|
||||||
@ -19,13 +18,13 @@ def _try_loading_included_file_formats() -> Dict[str, Type[BaseReader]]:
|
|||||||
from llama_index.readers.file.ipynb import IPYNBReader
|
from llama_index.readers.file.ipynb import IPYNBReader
|
||||||
from llama_index.readers.file.markdown import MarkdownReader
|
from llama_index.readers.file.markdown import MarkdownReader
|
||||||
from llama_index.readers.file.mbox import MboxReader
|
from llama_index.readers.file.mbox import MboxReader
|
||||||
from llama_index.readers.file.tabular import PandasCSVReader
|
|
||||||
from llama_index.readers.file.slides import PptxReader
|
from llama_index.readers.file.slides import PptxReader
|
||||||
|
from llama_index.readers.file.tabular import PandasCSVReader
|
||||||
from llama_index.readers.file.video_audio import VideoAudioReader
|
from llama_index.readers.file.video_audio import VideoAudioReader
|
||||||
except ImportError:
|
except ImportError as e:
|
||||||
raise ImportError("`llama-index-readers-file` package not found")
|
raise ImportError("`llama-index-readers-file` package not found") from e
|
||||||
|
|
||||||
default_file_reader_cls: Dict[str, Type[BaseReader]] = {
|
default_file_reader_cls: dict[str, type[BaseReader]] = {
|
||||||
".hwp": HWPReader,
|
".hwp": HWPReader,
|
||||||
".pdf": PDFReader,
|
".pdf": PDFReader,
|
||||||
".docx": DocxReader,
|
".docx": DocxReader,
|
||||||
|
@ -2,15 +2,14 @@ import logging
|
|||||||
|
|
||||||
from injector import inject, singleton
|
from injector import inject, singleton
|
||||||
from llama_index.core.llms import LLM, MockLLM
|
from llama_index.core.llms import LLM, MockLLM
|
||||||
from llama_index.core.utils import set_global_tokenizer
|
|
||||||
from llama_index.core.settings import Settings as LlamaIndexSettings
|
from llama_index.core.settings import Settings as LlamaIndexSettings
|
||||||
|
from llama_index.core.utils import set_global_tokenizer
|
||||||
from transformers import AutoTokenizer # type: ignore
|
from transformers import AutoTokenizer # type: ignore
|
||||||
|
|
||||||
from private_gpt.components.llm.prompt_helper import get_prompt_style
|
from private_gpt.components.llm.prompt_helper import get_prompt_style
|
||||||
from private_gpt.paths import models_cache_path, models_path
|
from private_gpt.paths import models_cache_path, models_path
|
||||||
from private_gpt.settings.settings import Settings
|
from private_gpt.settings.settings import Settings
|
||||||
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
@ -60,12 +60,14 @@ class VectorStoreComponent:
|
|||||||
|
|
||||||
case "chroma":
|
case "chroma":
|
||||||
try:
|
try:
|
||||||
from private_gpt.components.vector_store.batched_chroma import \
|
|
||||||
BatchedChromaVectorStore
|
|
||||||
import chromadb # type: ignore
|
import chromadb # type: ignore
|
||||||
from chromadb.config import ( # type: ignore
|
from chromadb.config import ( # type: ignore
|
||||||
Settings as ChromaSettings,
|
Settings as ChromaSettings,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
from private_gpt.components.vector_store.batched_chroma import (
|
||||||
|
BatchedChromaVectorStore,
|
||||||
|
)
|
||||||
except ImportError as e:
|
except ImportError as e:
|
||||||
raise ImportError(
|
raise ImportError(
|
||||||
"ChromaDB dependencies not found, install with `poetry install --extras chroma`"
|
"ChromaDB dependencies not found, install with `poetry install --extras chroma`"
|
||||||
|
@ -6,6 +6,7 @@ from fastapi.middleware.cors import CORSMiddleware
|
|||||||
from injector import Injector
|
from injector import Injector
|
||||||
from llama_index.core.callbacks import CallbackManager
|
from llama_index.core.callbacks import CallbackManager
|
||||||
from llama_index.core.callbacks.global_handlers import create_global_handler
|
from llama_index.core.callbacks.global_handlers import create_global_handler
|
||||||
|
from llama_index.core.settings import Settings as LlamaIndexSettings
|
||||||
|
|
||||||
from private_gpt.server.chat.chat_router import chat_router
|
from private_gpt.server.chat.chat_router import chat_router
|
||||||
from private_gpt.server.chunks.chunks_router import chunks_router
|
from private_gpt.server.chunks.chunks_router import chunks_router
|
||||||
@ -14,7 +15,6 @@ from private_gpt.server.embeddings.embeddings_router import embeddings_router
|
|||||||
from private_gpt.server.health.health_router import health_router
|
from private_gpt.server.health.health_router import health_router
|
||||||
from private_gpt.server.ingest.ingest_router import ingest_router
|
from private_gpt.server.ingest.ingest_router import ingest_router
|
||||||
from private_gpt.settings.settings import Settings
|
from private_gpt.settings.settings import Settings
|
||||||
from llama_index.core.settings import Settings as LlamaIndexSettings
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -105,6 +105,7 @@ class ChatService:
|
|||||||
# TODO ContextChatEngine is still not migrated by LlamaIndex to accept
|
# TODO ContextChatEngine is still not migrated by LlamaIndex to accept
|
||||||
# llm directly, so we are passing legacy ServiceContext until it is fixed.
|
# llm directly, so we are passing legacy ServiceContext until it is fixed.
|
||||||
from llama_index.core import ServiceContext
|
from llama_index.core import ServiceContext
|
||||||
|
|
||||||
return ContextChatEngine.from_defaults(
|
return ContextChatEngine.from_defaults(
|
||||||
system_prompt=system_prompt,
|
system_prompt=system_prompt,
|
||||||
retriever=vector_index_retriever,
|
retriever=vector_index_retriever,
|
||||||
|
Loading…
Reference in New Issue
Block a user