mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-08-04 01:50:08 +00:00
rebuild params
This commit is contained in:
parent
f0e17ed8f1
commit
eef244fe92
@ -33,6 +33,13 @@ class PromptRequest(BaseModel):
|
|||||||
stop: Optional[List[str]] = None
|
stop: Optional[List[str]] = None
|
||||||
|
|
||||||
|
|
||||||
|
class StreamRequest(BaseModel):
|
||||||
|
model: str
|
||||||
|
prompt: str
|
||||||
|
temperature: float
|
||||||
|
max_new_tokens: int
|
||||||
|
stop: str
|
||||||
|
|
||||||
class EmbeddingRequest(BaseModel):
|
class EmbeddingRequest(BaseModel):
|
||||||
prompt: str
|
prompt: str
|
||||||
|
|
||||||
@ -64,10 +71,16 @@ def generate_stream_gate(params):
|
|||||||
|
|
||||||
|
|
||||||
@app.post("/generate_stream")
|
@app.post("/generate_stream")
|
||||||
def api_generate_stream(request: Request):
|
def api_generate_stream(request: StreamRequest):
|
||||||
global model_semaphore, global_counter
|
global model_semaphore, global_counter
|
||||||
global_counter += 1
|
global_counter += 1
|
||||||
params = request.json()
|
params = {
|
||||||
|
"prompt": request.prompt,
|
||||||
|
"model": request.model,
|
||||||
|
"temperature": request.temperature,
|
||||||
|
"max_new_tokens": request.max_new_tokens,
|
||||||
|
"stop": request.stop
|
||||||
|
}
|
||||||
print(model, tokenizer, params, DEVICE)
|
print(model, tokenizer, params, DEVICE)
|
||||||
# if model_semaphore is None:
|
# if model_semaphore is None:
|
||||||
# model_semaphore = asyncio.Semaphore(LIMIT_MODEL_CONCURRENCY)
|
# model_semaphore = asyncio.Semaphore(LIMIT_MODEL_CONCURRENCY)
|
||||||
|
Loading…
Reference in New Issue
Block a user