mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-08-11 20:32:10 +00:00
Revise code for zero errors in make check/test
This commit is contained in:
parent
d737cb8d1c
commit
987381247f
@ -184,19 +184,28 @@ class PrivateGptUi:
|
|||||||
docs_ids.append(ingested_document.doc_id)
|
docs_ids.append(ingested_document.doc_id)
|
||||||
context_filter = ContextFilter(docs_ids=docs_ids)
|
context_filter = ContextFilter(docs_ids=docs_ids)
|
||||||
|
|
||||||
query_stream = self._chat_service.stream_chat(
|
if self._response_style:
|
||||||
all_messages, use_context=False
|
query_stream = self._chat_service.stream_chat(
|
||||||
) if self._response_style else self._chat_service.chat(
|
all_messages, use_context=False
|
||||||
all_messages, use_context=False
|
)
|
||||||
)
|
yield from yield_deltas(query_stream)
|
||||||
yield from (yield_deltas(query_stream) if self._response_style else [query_stream.response])
|
else:
|
||||||
|
query_response = self._chat_service.chat(
|
||||||
|
all_messages, use_context=False
|
||||||
|
).response
|
||||||
|
yield from [query_response]
|
||||||
|
|
||||||
case Modes.BASIC_CHAT_MODE:
|
case Modes.BASIC_CHAT_MODE:
|
||||||
llm_stream = self._chat_service.stream_chat(
|
if self._response_style:
|
||||||
all_messages, use_context=False
|
llm_stream = self._chat_service.stream_chat(
|
||||||
) if self._response_style else self._chat_service.chat(
|
all_messages, use_context=False
|
||||||
all_messages, use_context=False
|
)
|
||||||
)
|
yield from yield_deltas(llm_stream)
|
||||||
yield from (yield_deltas(llm_stream) if self._response_style else [llm_stream.response])
|
else:
|
||||||
|
llm_response = self._chat_service.chat(
|
||||||
|
all_messages, use_context=False
|
||||||
|
).response
|
||||||
|
yield from [llm_response]
|
||||||
|
|
||||||
case Modes.SEARCH_MODE:
|
case Modes.SEARCH_MODE:
|
||||||
response = self._chunks_service.retrieve_relevant(
|
response = self._chunks_service.retrieve_relevant(
|
||||||
@ -224,20 +233,20 @@ class PrivateGptUi:
|
|||||||
docs_ids.append(ingested_document.doc_id)
|
docs_ids.append(ingested_document.doc_id)
|
||||||
context_filter = ContextFilter(docs_ids=docs_ids)
|
context_filter = ContextFilter(docs_ids=docs_ids)
|
||||||
|
|
||||||
summary_stream = self._summarize_service.stream_summarize(
|
if self._response_style:
|
||||||
use_context=True,
|
summary_stream = self._summarize_service.stream_summarize(
|
||||||
context_filter=context_filter,
|
use_context=True,
|
||||||
instructions=message,
|
context_filter=context_filter,
|
||||||
)
|
instructions=message,
|
||||||
yield from yield_tokens(summary_stream)
|
)
|
||||||
'''
|
yield from yield_tokens(summary_stream)
|
||||||
summary_stream = self._summarize_service.summarize_stream(
|
else:
|
||||||
all_messages, use_context=False
|
summary_response = self._summarize_service.summarize(
|
||||||
) if self._response_style else self._summarize_service.summarize(
|
use_context=True,
|
||||||
all_messages, use_context=False
|
context_filter=context_filter,
|
||||||
)
|
instructions=message,
|
||||||
yield from yield_tokens(summary_stream) if response_style else summary_stream
|
)
|
||||||
'''
|
yield from summary_response
|
||||||
|
|
||||||
# On initialization and on mode change, this function set the system prompt
|
# On initialization and on mode change, this function set the system prompt
|
||||||
# to the default prompt based on the mode (and user settings).
|
# to the default prompt based on the mode (and user settings).
|
||||||
@ -290,7 +299,7 @@ class PrivateGptUi:
|
|||||||
gr.update(value=self._explanation_mode),
|
gr.update(value=self._explanation_mode),
|
||||||
]
|
]
|
||||||
|
|
||||||
def _set_response_style(self, response_style: str) -> None:
|
def _set_response_style(self, response_style: bool) -> None:
|
||||||
self._response_style = response_style
|
self._response_style = response_style
|
||||||
|
|
||||||
def _list_ingested_files(self) -> list[list[str]]:
|
def _list_ingested_files(self) -> list[list[str]]:
|
||||||
|
Loading…
Reference in New Issue
Block a user