mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-07-23 12:21:08 +00:00
fix:core agent dependent prompt service bug (#2129)
This commit is contained in:
parent
2decf02c12
commit
b392d51adf
@ -22,7 +22,6 @@ from dbgpt.core.interface.message import ModelMessageRoleType
|
||||
|
||||
# TODO: Don't dependent on MixinLLMOperator
|
||||
from dbgpt.model.operators.llm_operator import MixinLLMOperator
|
||||
from dbgpt.serve.prompt.api.endpoints import get_service
|
||||
from dbgpt.util.i18n_utils import _
|
||||
|
||||
from .... import ActionOutput
|
||||
@ -291,6 +290,7 @@ class AWELAgentOperator(
|
||||
|
||||
prompt_template = None
|
||||
if self.awel_agent.agent_prompt:
|
||||
from dbgpt.serve.prompt.api.endpoints import get_service
|
||||
prompt_service = get_service()
|
||||
prompt_template = prompt_service.get_template(
|
||||
self.awel_agent.agent_prompt.code
|
||||
|
@ -12,8 +12,6 @@ from dbgpt.core.awel.flow import (
|
||||
ResourceCategory,
|
||||
register_resource,
|
||||
)
|
||||
from dbgpt.serve.prompt.api.endpoints import get_service
|
||||
|
||||
from ....resource.base import AgentResource, ResourceType
|
||||
from ....resource.manage import get_resource_manager
|
||||
from ....util.llm.llm import LLMConfig, LLMStrategyType
|
||||
@ -21,6 +19,7 @@ from ...agent_manage import get_agent_manager
|
||||
|
||||
|
||||
def _agent_resource_prompt_values() -> List[OptionValue]:
|
||||
from dbgpt.serve.prompt.api.endpoints import get_service
|
||||
prompt_service = get_service()
|
||||
prompts = prompt_service.get_target_prompt()
|
||||
return [
|
||||
|
Loading…
Reference in New Issue
Block a user