feat: structured prompt data interaction between dbgpt-server and llm-server

This commit is contained in:
FangYin Cheng
2023-07-18 18:07:09 +08:00
parent f5d817f9db
commit d805e6ab42
5 changed files with 151 additions and 79 deletions

View File

@@ -5,6 +5,7 @@ import asyncio
import json
import os
import sys
from typing import List
import uvicorn
from fastapi import BackgroundTasks, FastAPI, Request
@@ -24,6 +25,7 @@ from pilot.configs.model_config import *
from pilot.model.llm_out.vicuna_base_llm import get_embeddings
from pilot.model.loader import ModelLoader
from pilot.server.chat_adapter import get_llm_chat_adapter
from pilot.scene.base_message import ModelMessage
CFG = Config()
@@ -128,6 +130,7 @@ app = FastAPI()
class PromptRequest(BaseModel):
messages: List[ModelMessage]
prompt: str
temperature: float
max_new_tokens: int
@@ -170,6 +173,7 @@ async def api_generate_stream(request: Request):
@app.post("/generate")
def generate(prompt_request: PromptRequest) -> str:
params = {
"messages": prompt_request.messages,
"prompt": prompt_request.prompt,
"temperature": prompt_request.temperature,
"max_new_tokens": prompt_request.max_new_tokens,