mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-09-17 15:10:14 +00:00
feat: structured prompt data interaction between dbgpt-server and llm-server
This commit is contained in:
@@ -5,6 +5,7 @@ import asyncio
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
from typing import List
|
||||
|
||||
import uvicorn
|
||||
from fastapi import BackgroundTasks, FastAPI, Request
|
||||
@@ -24,6 +25,7 @@ from pilot.configs.model_config import *
|
||||
from pilot.model.llm_out.vicuna_base_llm import get_embeddings
|
||||
from pilot.model.loader import ModelLoader
|
||||
from pilot.server.chat_adapter import get_llm_chat_adapter
|
||||
from pilot.scene.base_message import ModelMessage
|
||||
|
||||
CFG = Config()
|
||||
|
||||
@@ -128,6 +130,7 @@ app = FastAPI()
|
||||
|
||||
|
||||
class PromptRequest(BaseModel):
|
||||
messages: List[ModelMessage]
|
||||
prompt: str
|
||||
temperature: float
|
||||
max_new_tokens: int
|
||||
@@ -170,6 +173,7 @@ async def api_generate_stream(request: Request):
|
||||
@app.post("/generate")
|
||||
def generate(prompt_request: PromptRequest) -> str:
|
||||
params = {
|
||||
"messages": prompt_request.messages,
|
||||
"prompt": prompt_request.prompt,
|
||||
"temperature": prompt_request.temperature,
|
||||
"max_new_tokens": prompt_request.max_new_tokens,
|
||||
|
Reference in New Issue
Block a user