mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-07-03 02:27:25 +00:00
Update files for ruff
This commit is contained in:
parent
6bb4d18ade
commit
eab8799174
@ -7,9 +7,11 @@ from llama_index.core.chat_engine.types import (
|
|||||||
)
|
)
|
||||||
from llama_index.core.indices import VectorStoreIndex
|
from llama_index.core.indices import VectorStoreIndex
|
||||||
from llama_index.core.indices.postprocessor import MetadataReplacementPostProcessor
|
from llama_index.core.indices.postprocessor import MetadataReplacementPostProcessor
|
||||||
from llama_index.core.postprocessor import SimilarityPostprocessor
|
|
||||||
from llama_index.core.postprocessor import KeywordNodePostprocessor
|
|
||||||
from llama_index.core.llms import ChatMessage, MessageRole
|
from llama_index.core.llms import ChatMessage, MessageRole
|
||||||
|
from llama_index.core.postprocessor import (
|
||||||
|
KeywordNodePostprocessor,
|
||||||
|
SimilarityPostprocessor,
|
||||||
|
)
|
||||||
from llama_index.core.storage import StorageContext
|
from llama_index.core.storage import StorageContext
|
||||||
from llama_index.core.types import TokenGen
|
from llama_index.core.types import TokenGen
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
@ -24,6 +26,7 @@ from private_gpt.open_ai.extensions.context_filter import ContextFilter
|
|||||||
from private_gpt.server.chunks.chunks_service import Chunk
|
from private_gpt.server.chunks.chunks_service import Chunk
|
||||||
from private_gpt.settings.settings import settings
|
from private_gpt.settings.settings import settings
|
||||||
|
|
||||||
|
|
||||||
class Completion(BaseModel):
|
class Completion(BaseModel):
|
||||||
response: str
|
response: str
|
||||||
sources: list[Chunk] | None = None
|
sources: list[Chunk] | None = None
|
||||||
@ -114,8 +117,13 @@ class ChatService:
|
|||||||
llm=self.llm_component.llm, # Takes no effect at the moment
|
llm=self.llm_component.llm, # Takes no effect at the moment
|
||||||
node_postprocessors=[
|
node_postprocessors=[
|
||||||
MetadataReplacementPostProcessor(target_metadata_key="window"),
|
MetadataReplacementPostProcessor(target_metadata_key="window"),
|
||||||
SimilarityPostprocessor(similarity_cutoff=settings().llm.similarity_value),
|
SimilarityPostprocessor(
|
||||||
KeywordNodePostprocessor(required_keywords=settings().llm.keywords_include, exclude_keywords=settings().llm.keywords_exclude),
|
similarity_cutoff=settings().llm.similarity_value
|
||||||
|
),
|
||||||
|
KeywordNodePostprocessor(
|
||||||
|
required_keywords=settings().llm.keywords_include,
|
||||||
|
exclude_keywords=settings().llm.keywords_exclude,
|
||||||
|
),
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
@ -125,7 +133,9 @@ class ChatService:
|
|||||||
llm=self.llm_component.llm, # Takes no effect at the moment
|
llm=self.llm_component.llm, # Takes no effect at the moment
|
||||||
node_postprocessors=[
|
node_postprocessors=[
|
||||||
MetadataReplacementPostProcessor(target_metadata_key="window"),
|
MetadataReplacementPostProcessor(target_metadata_key="window"),
|
||||||
SimilarityPostprocessor(similarity_cutoff=settings().llm.similarity_value),
|
SimilarityPostprocessor(
|
||||||
|
similarity_cutoff=settings().llm.similarity_value
|
||||||
|
),
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user