mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-09-01 09:06:55 +00:00
fix: agent examples
This commit is contained in:
@@ -6,8 +6,8 @@
|
|||||||
Set env params.
|
Set env params.
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
export OPENAI_API_KEY=sk-xx
|
export SILICONFLOW_API_KEY=sk-xx
|
||||||
export OPENAI_API_BASE=https://xx:80/v1
|
export SILICONFLOW_API_BASE=https://xx:80/v1
|
||||||
|
|
||||||
run example.
|
run example.
|
||||||
..code-block:: shell
|
..code-block:: shell
|
||||||
@@ -15,6 +15,7 @@
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import os
|
||||||
|
|
||||||
from dbgpt.agent import (
|
from dbgpt.agent import (
|
||||||
AgentContext,
|
AgentContext,
|
||||||
@@ -32,12 +33,14 @@ initialize_tracer(
|
|||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(
|
llm_client = SiliconFlowLLMClient(
|
||||||
model_alias="gpt-4o",
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
context: AgentContext = AgentContext(
|
context: AgentContext = AgentContext(
|
||||||
|
@@ -6,14 +6,13 @@
|
|||||||
Set env params.
|
Set env params.
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
export OPENAI_API_KEY=sk-xx
|
export SILICONFLOW_API_KEY=sk-xx
|
||||||
export OPENAI_API_BASE=https://xx:80/v1
|
export SILICONFLOW_API_BASE=https://xx:80/v1
|
||||||
|
|
||||||
run example.
|
run example.
|
||||||
..code-block:: shell
|
..code-block:: shell
|
||||||
python examples/agents/awel_layout_agents_chat_examples.py
|
python examples/agents/awel_layout_agents_chat_examples.py
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
import os
|
import os
|
||||||
|
|
||||||
@@ -34,15 +33,16 @@ initialize_tracer("/tmp/agent_trace.jsonl", create_system_app=True)
|
|||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
agent_memory.gpts_memory.init(conv_id="test456")
|
agent_memory.gpts_memory.init(conv_id="test456")
|
||||||
try:
|
try:
|
||||||
from dbgpt.model.proxy.llms.tongyi import TongyiLLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = TongyiLLMClient(
|
llm_client = SiliconFlowLLMClient(
|
||||||
model_alias="qwen2-72b-instruct",
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
context: AgentContext = AgentContext(conv_id="test456", gpts_app_name="信息析助手")
|
context: AgentContext = AgentContext(conv_id="test456", gpts_app_name="信息析助手")
|
||||||
|
@@ -44,12 +44,17 @@ def count_directory_files(path: Annotated[str, Doc("The directory path")]) -> in
|
|||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
|
||||||
context: AgentContext = AgentContext(conv_id="test456")
|
|
||||||
|
|
||||||
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="test456")
|
||||||
|
|
||||||
|
context: AgentContext = AgentContext(conv_id="test456", gpts_app_name="工具助手")
|
||||||
|
|
||||||
tools = ToolPack([simple_calculator, count_directory_files])
|
tools = ToolPack([simple_calculator, count_directory_files])
|
||||||
|
|
||||||
@@ -77,7 +82,7 @@ async def main():
|
|||||||
)
|
)
|
||||||
|
|
||||||
# dbgpt-vis message infos
|
# dbgpt-vis message infos
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("test456"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("test456"))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@@ -6,8 +6,8 @@
|
|||||||
Set env params.
|
Set env params.
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
export OPENAI_API_KEY=sk-xx
|
export SILICONFLOW_API_KEY=sk-xx
|
||||||
export OPENAI_API_BASE=https://xx:80/v1
|
export SILICONFLOW_API_BASE=https://xx:80/v1
|
||||||
|
|
||||||
run example.
|
run example.
|
||||||
..code-block:: shell
|
..code-block:: shell
|
||||||
@@ -20,19 +20,24 @@ import os
|
|||||||
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
||||||
from dbgpt.agent.expand.tool_assistant_agent import ToolAssistantAgent
|
from dbgpt.agent.expand.tool_assistant_agent import ToolAssistantAgent
|
||||||
from dbgpt.agent.resource import AutoGPTPluginToolPack
|
from dbgpt.agent.resource import AutoGPTPluginToolPack
|
||||||
|
from dbgpt.configs.model_config import ROOT_PATH
|
||||||
|
|
||||||
current_dir = os.getcwd()
|
test_plugin_dir = os.path.join(ROOT_PATH, "examples/test_files/plugins")
|
||||||
parent_dir = os.path.dirname(current_dir)
|
|
||||||
test_plugin_dir = os.path.join(parent_dir, "test_files/plugins")
|
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
llm_client = SiliconFlowLLMClient(
|
||||||
context: AgentContext = AgentContext(conv_id="test456")
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="test456")
|
||||||
|
|
||||||
|
context: AgentContext = AgentContext(conv_id="test456", gpts_app_name="插件对话助手")
|
||||||
|
|
||||||
tools = AutoGPTPluginToolPack(test_plugin_dir)
|
tools = AutoGPTPluginToolPack(test_plugin_dir)
|
||||||
|
|
||||||
@@ -54,7 +59,7 @@ async def main():
|
|||||||
)
|
)
|
||||||
|
|
||||||
# dbgpt-vis message infos
|
# dbgpt-vis message infos
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("test456"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("test456"))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@@ -6,8 +6,8 @@
|
|||||||
Set env params.
|
Set env params.
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
export OPENAI_API_KEY=sk-xx
|
export SILICONFLOW_API_KEY=sk-xx
|
||||||
export OPENAI_API_BASE=https://xx:80/v1
|
export SILICONFLOW_API_BASE=https://xx:80/v1
|
||||||
|
|
||||||
run example.
|
run example.
|
||||||
..code-block:: shell
|
..code-block:: shell
|
||||||
@@ -18,20 +18,27 @@ import asyncio
|
|||||||
import os
|
import os
|
||||||
|
|
||||||
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
||||||
from dbgpt.agent.expand.retrieve_summary_assistant_agent import (
|
from dbgpt.agent.expand.summary_assistant_agent import SummaryAssistantAgent
|
||||||
RetrieveSummaryAssistantAgent,
|
|
||||||
)
|
|
||||||
from dbgpt.configs.model_config import ROOT_PATH
|
from dbgpt.configs.model_config import ROOT_PATH
|
||||||
|
|
||||||
|
|
||||||
async def summary_example_with_success():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
context: AgentContext = AgentContext(
|
||||||
|
conv_id="retrieve_summarize", gpts_app_name="Summary Assistant"
|
||||||
|
)
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo-16k")
|
|
||||||
context: AgentContext = AgentContext(conv_id="retrieve_summarize")
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="retrieve_summarize")
|
||||||
|
|
||||||
summarizer = (
|
summarizer = (
|
||||||
await RetrieveSummaryAssistantAgent()
|
await SummaryAssistantAgent()
|
||||||
.bind(context)
|
.bind(context)
|
||||||
.bind(LLMConfig(llm_client=llm_client))
|
.bind(LLMConfig(llm_client=llm_client))
|
||||||
.bind(agent_memory)
|
.bind(agent_memory)
|
||||||
@@ -47,6 +54,7 @@ async def summary_example_with_success():
|
|||||||
"https://en.wikipedia.org/wiki/Chernobyl_disaster",
|
"https://en.wikipedia.org/wiki/Chernobyl_disaster",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
# TODO add a tool to load the pdf and internet files
|
||||||
await user_proxy.initiate_chat(
|
await user_proxy.initiate_chat(
|
||||||
recipient=summarizer,
|
recipient=summarizer,
|
||||||
reviewer=user_proxy,
|
reviewer=user_proxy,
|
||||||
@@ -55,11 +63,8 @@ async def summary_example_with_success():
|
|||||||
)
|
)
|
||||||
|
|
||||||
# dbgpt-vis message infos
|
# dbgpt-vis message infos
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("retrieve_summarize"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("retrieve_summarize"))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
asyncio.run(summary_example_with_success())
|
asyncio.run(main())
|
||||||
print(
|
|
||||||
"\033[92m=======================The Summary Assistant with Successful Results Ended==================\n\n\033[91m"
|
|
||||||
)
|
|
||||||
|
@@ -10,9 +10,9 @@ You can limit the memory and file system resources available to the code executi
|
|||||||
environment. The code execution environment is isolated from the host system,
|
environment. The code execution environment is isolated from the host system,
|
||||||
preventing access to the internet and other external resources.
|
preventing access to the internet and other external resources.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
|
import os
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
|
|
||||||
from dbgpt.agent import (
|
from dbgpt.agent import (
|
||||||
@@ -270,11 +270,28 @@ class SandboxCodeAssistantAgent(ConversableAgent):
|
|||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-4o-mini")
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
context: AgentContext = AgentContext(conv_id="test123")
|
context: AgentContext = AgentContext(conv_id="test123")
|
||||||
agent_memory = AgentMemory(HybridMemory[AgentMemoryFragment].from_chroma())
|
|
||||||
|
# TODO Embedding and Rerank model refactor
|
||||||
|
from dbgpt.rag.embedding import OpenAPIEmbeddings
|
||||||
|
|
||||||
|
silicon_embeddings = OpenAPIEmbeddings(
|
||||||
|
api_url=os.getenv("SILICONFLOW_API_BASE") + "/embeddings",
|
||||||
|
api_key=os.getenv("SILICONFLOW_API_KEY"),
|
||||||
|
model_name="BAAI/bge-large-zh-v1.5",
|
||||||
|
)
|
||||||
|
agent_memory = AgentMemory(
|
||||||
|
HybridMemory[AgentMemoryFragment].from_chroma(
|
||||||
|
embeddings=silicon_embeddings,
|
||||||
|
)
|
||||||
|
)
|
||||||
agent_memory.gpts_memory.init("test123")
|
agent_memory.gpts_memory.init("test123")
|
||||||
|
|
||||||
coder = (
|
coder = (
|
||||||
|
@@ -17,28 +17,21 @@
|
|||||||
import asyncio
|
import asyncio
|
||||||
import os
|
import os
|
||||||
|
|
||||||
from dbgpt.agent import (
|
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
||||||
AgentContext,
|
|
||||||
AgentMemory,
|
|
||||||
AgentMemoryFragment,
|
|
||||||
HybridMemory,
|
|
||||||
LLMConfig,
|
|
||||||
UserProxyAgent,
|
|
||||||
)
|
|
||||||
from dbgpt.agent.expand.code_assistant_agent import CodeAssistantAgent
|
from dbgpt.agent.expand.code_assistant_agent import CodeAssistantAgent
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
# llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
llm_client = SiliconFlowLLMClient(
|
||||||
from dbgpt.model.proxy.llms.tongyi import TongyiLLMClient
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
llm_client = TongyiLLMClient(
|
),
|
||||||
model_alias="qwen2-72b-instruct",
|
|
||||||
)
|
)
|
||||||
|
|
||||||
context: AgentContext = AgentContext(conv_id="test123")
|
context: AgentContext = AgentContext(conv_id="test123", gpts_app_name="代码助手")
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
agent_memory.gpts_memory.init(conv_id="test123")
|
agent_memory.gpts_memory.init(conv_id="test123")
|
||||||
try:
|
try:
|
||||||
|
@@ -15,18 +15,25 @@
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import os
|
||||||
|
|
||||||
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
||||||
from dbgpt.agent.expand.summary_assistant_agent import SummaryAssistantAgent
|
from dbgpt.agent.expand.summary_assistant_agent import SummaryAssistantAgent
|
||||||
|
|
||||||
|
|
||||||
async def summary_example_with_success():
|
async def summary_example_with_success():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
context: AgentContext = AgentContext(conv_id="summarize")
|
context: AgentContext = AgentContext(conv_id="summarize")
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="summarize")
|
||||||
|
|
||||||
summarizer = (
|
summarizer = (
|
||||||
await SummaryAssistantAgent()
|
await SummaryAssistantAgent()
|
||||||
.bind(context)
|
.bind(context)
|
||||||
@@ -71,16 +78,22 @@ async def summary_example_with_success():
|
|||||||
)
|
)
|
||||||
|
|
||||||
# dbgpt-vis message infos
|
# dbgpt-vis message infos
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("summarize"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("summarize"))
|
||||||
|
|
||||||
|
|
||||||
async def summary_example_with_faliure():
|
async def summary_example_with_faliure():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
context: AgentContext = AgentContext(conv_id="summarize")
|
context: AgentContext = AgentContext(conv_id="summarize")
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="summarize")
|
||||||
|
|
||||||
summarizer = (
|
summarizer = (
|
||||||
await SummaryAssistantAgent()
|
await SummaryAssistantAgent()
|
||||||
.bind(context)
|
.bind(context)
|
||||||
@@ -92,7 +105,6 @@ async def summary_example_with_faliure():
|
|||||||
user_proxy = await UserProxyAgent().bind(agent_memory).bind(context).build()
|
user_proxy = await UserProxyAgent().bind(agent_memory).bind(context).build()
|
||||||
|
|
||||||
# Test the failure example
|
# Test the failure example
|
||||||
|
|
||||||
await user_proxy.initiate_chat(
|
await user_proxy.initiate_chat(
|
||||||
recipient=summarizer,
|
recipient=summarizer,
|
||||||
reviewer=user_proxy,
|
reviewer=user_proxy,
|
||||||
@@ -110,7 +122,7 @@ async def summary_example_with_faliure():
|
|||||||
""",
|
""",
|
||||||
)
|
)
|
||||||
|
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("summarize"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("summarize"))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@@ -20,24 +20,28 @@ import os
|
|||||||
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
from dbgpt.agent import AgentContext, AgentMemory, LLMConfig, UserProxyAgent
|
||||||
from dbgpt.agent.expand.data_scientist_agent import DataScientistAgent
|
from dbgpt.agent.expand.data_scientist_agent import DataScientistAgent
|
||||||
from dbgpt.agent.resource import SQLiteDBResource
|
from dbgpt.agent.resource import SQLiteDBResource
|
||||||
|
from dbgpt.configs.model_config import ROOT_PATH
|
||||||
from dbgpt.util.tracer import initialize_tracer
|
from dbgpt.util.tracer import initialize_tracer
|
||||||
|
|
||||||
current_dir = os.getcwd()
|
test_plugin_dir = os.path.join(ROOT_PATH, "test_files")
|
||||||
parent_dir = os.path.dirname(current_dir)
|
|
||||||
test_plugin_dir = os.path.join(parent_dir, "test_files")
|
|
||||||
|
|
||||||
initialize_tracer("/tmp/agent_trace.jsonl", create_system_app=True)
|
initialize_tracer("/tmp/agent_trace.jsonl", create_system_app=True)
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
from dbgpt.model.proxy import OpenAILLMClient
|
from dbgpt.model.proxy.llms.siliconflow import SiliconFlowLLMClient
|
||||||
|
|
||||||
llm_client = OpenAILLMClient(model_alias="gpt-3.5-turbo")
|
llm_client = SiliconFlowLLMClient(
|
||||||
|
model_alias=os.getenv(
|
||||||
|
"SILICONFLOW_MODEL_VERSION", "Qwen/Qwen2.5-Coder-32B-Instruct"
|
||||||
|
),
|
||||||
|
)
|
||||||
context: AgentContext = AgentContext(conv_id="test456")
|
context: AgentContext = AgentContext(conv_id="test456")
|
||||||
|
|
||||||
agent_memory = AgentMemory()
|
agent_memory = AgentMemory()
|
||||||
|
agent_memory.gpts_memory.init(conv_id="test456")
|
||||||
|
|
||||||
sqlite_resource = SQLiteDBResource("SQLite Database", f"{test_plugin_dir}/dbgpt.db")
|
sqlite_resource = SQLiteDBResource("SQLite Database", f"{test_plugin_dir}/dbgpt.db")
|
||||||
|
|
||||||
user_proxy = await UserProxyAgent().bind(agent_memory).bind(context).build()
|
user_proxy = await UserProxyAgent().bind(agent_memory).bind(context).build()
|
||||||
|
|
||||||
sql_boy = (
|
sql_boy = (
|
||||||
@@ -56,7 +60,7 @@ async def main():
|
|||||||
)
|
)
|
||||||
|
|
||||||
## dbgpt-vis message infos
|
## dbgpt-vis message infos
|
||||||
print(await agent_memory.gpts_memory.one_chat_completions("test456"))
|
print(await agent_memory.gpts_memory.app_link_chat_message("test456"))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
Reference in New Issue
Block a user