mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-08-20 01:07:15 +00:00
refactor: webserver chat delete temperature, max_new_tokens
This commit is contained in:
parent
9137e413b6
commit
8baa1d9464
@ -33,9 +33,6 @@ class SourceEmbedding(ABC):
|
|||||||
self.vector_store_config = vector_store_config
|
self.vector_store_config = vector_store_config
|
||||||
self.embedding_args = embedding_args
|
self.embedding_args = embedding_args
|
||||||
self.embeddings = vector_store_config["embeddings"]
|
self.embeddings = vector_store_config["embeddings"]
|
||||||
self.vector_client = VectorStoreConnector(
|
|
||||||
CFG.VECTOR_STORE_TYPE, vector_store_config
|
|
||||||
)
|
|
||||||
|
|
||||||
@abstractmethod
|
@abstractmethod
|
||||||
@register
|
@register
|
||||||
@ -59,11 +56,17 @@ class SourceEmbedding(ABC):
|
|||||||
@register
|
@register
|
||||||
def index_to_store(self, docs):
|
def index_to_store(self, docs):
|
||||||
"""index to vector store"""
|
"""index to vector store"""
|
||||||
return self.vector_client.load_document(docs)
|
self.vector_client = VectorStoreConnector(
|
||||||
|
CFG.VECTOR_STORE_TYPE, self.vector_store_config
|
||||||
|
)
|
||||||
|
self.vector_client.load_document(docs)
|
||||||
|
|
||||||
@register
|
@register
|
||||||
def similar_search(self, doc, topk):
|
def similar_search(self, doc, topk):
|
||||||
"""vector store similarity_search"""
|
"""vector store similarity_search"""
|
||||||
|
self.vector_client = VectorStoreConnector(
|
||||||
|
CFG.VECTOR_STORE_TYPE, self.vector_store_config
|
||||||
|
)
|
||||||
try:
|
try:
|
||||||
ans = self.vector_client.similar_search(doc, topk)
|
ans = self.vector_client.similar_search(doc, topk)
|
||||||
except NotEnoughElementsException:
|
except NotEnoughElementsException:
|
||||||
@ -71,6 +74,9 @@ class SourceEmbedding(ABC):
|
|||||||
return ans
|
return ans
|
||||||
|
|
||||||
def vector_name_exist(self):
|
def vector_name_exist(self):
|
||||||
|
self.vector_client = VectorStoreConnector(
|
||||||
|
CFG.VECTOR_STORE_TYPE, self.vector_store_config
|
||||||
|
)
|
||||||
return self.vector_client.vector_name_exists()
|
return self.vector_client.vector_name_exists()
|
||||||
|
|
||||||
def source_embedding(self):
|
def source_embedding(self):
|
||||||
|
@ -297,39 +297,29 @@ def http_bot(
|
|||||||
|
|
||||||
if ChatScene.ChatWithDbExecute == scene:
|
if ChatScene.ChatWithDbExecute == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"db_name": db_selector,
|
"db_name": db_selector,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
}
|
}
|
||||||
elif ChatScene.ChatWithDbQA == scene:
|
elif ChatScene.ChatWithDbQA == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"db_name": db_selector,
|
"db_name": db_selector,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
}
|
}
|
||||||
elif ChatScene.ChatExecution == scene:
|
elif ChatScene.ChatExecution == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"plugin_selector": plugin_selector,
|
"plugin_selector": plugin_selector,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
}
|
}
|
||||||
elif ChatScene.ChatNormal == scene:
|
elif ChatScene.ChatNormal == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
}
|
}
|
||||||
elif ChatScene.ChatDefaultKnowledge == scene:
|
elif ChatScene.ChatDefaultKnowledge == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
}
|
}
|
||||||
@ -341,8 +331,6 @@ def http_bot(
|
|||||||
}
|
}
|
||||||
elif ChatScene.ChatUrlKnowledge == scene:
|
elif ChatScene.ChatUrlKnowledge == scene:
|
||||||
chat_param = {
|
chat_param = {
|
||||||
"temperature": temperature,
|
|
||||||
"max_new_tokens": max_new_tokens,
|
|
||||||
"chat_session_id": state.conv_id,
|
"chat_session_id": state.conv_id,
|
||||||
"user_input": state.last_user_input,
|
"user_input": state.last_user_input,
|
||||||
"url": url_input,
|
"url": url_input,
|
||||||
|
@ -49,6 +49,7 @@ class LocalKnowledgeInit:
|
|||||||
space = KnowledgeSpaceRequest
|
space = KnowledgeSpaceRequest
|
||||||
space.name = self.vector_store_config["vector_store_name"]
|
space.name = self.vector_store_config["vector_store_name"]
|
||||||
space.desc = "knowledge_init.py"
|
space.desc = "knowledge_init.py"
|
||||||
|
space.vector_type = CFG.VECTOR_STORE_TYPE
|
||||||
space.owner = "DB-GPT"
|
space.owner = "DB-GPT"
|
||||||
knowledge_space_service.create_knowledge_space(space)
|
knowledge_space_service.create_knowledge_space(space)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user