mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-08-10 12:42:34 +00:00
add debug log
This commit is contained in:
parent
65ead15b65
commit
905f14cf7c
@ -48,8 +48,8 @@ def generate_stream_gate(params):
|
|||||||
params,
|
params,
|
||||||
DEVICE,
|
DEVICE,
|
||||||
MAX_POSITION_EMBEDDINGS,
|
MAX_POSITION_EMBEDDINGS,
|
||||||
2,
|
|
||||||
):
|
):
|
||||||
|
print("output: ", output)
|
||||||
ret = {
|
ret = {
|
||||||
"text": output,
|
"text": output,
|
||||||
"error_code": 0,
|
"error_code": 0,
|
||||||
@ -68,7 +68,7 @@ async def api_generate_stream(request: Request):
|
|||||||
global model_semaphore, global_counter
|
global model_semaphore, global_counter
|
||||||
global_counter += 1
|
global_counter += 1
|
||||||
params = await request.json()
|
params = await request.json()
|
||||||
|
print(model, tokenizer, params, DEVICE)
|
||||||
if model_semaphore is None:
|
if model_semaphore is None:
|
||||||
model_semaphore = asyncio.Semaphore(LIMIT_MODEL_CONCURRENCY)
|
model_semaphore = asyncio.Semaphore(LIMIT_MODEL_CONCURRENCY)
|
||||||
await model_semaphore.acquire()
|
await model_semaphore.acquire()
|
||||||
|
@ -156,12 +156,13 @@ def http_bot(state, temperature, max_new_tokens, request: gr.Request):
|
|||||||
prompt = state.get_prompt()
|
prompt = state.get_prompt()
|
||||||
skip_echo_len = compute_skip_echo_len(prompt)
|
skip_echo_len = compute_skip_echo_len(prompt)
|
||||||
|
|
||||||
|
logger.info(f"State: {state}")
|
||||||
payload = {
|
payload = {
|
||||||
"model": model_name,
|
"model": model_name,
|
||||||
"prompt": prompt,
|
"prompt": prompt,
|
||||||
"temperature": temperature,
|
"temperature": temperature,
|
||||||
"max_new_tokens": max_new_tokens,
|
"max_new_tokens": max_new_tokens,
|
||||||
"stop": state.sep if state.sep_style == SeparatorStyle.SINGLE else None,
|
"stop": state.sep,
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.info(f"Request: \n {payload}")
|
logger.info(f"Request: \n {payload}")
|
||||||
@ -179,6 +180,7 @@ def http_bot(state, temperature, max_new_tokens, request: gr.Request):
|
|||||||
for chunk in response.iter_lines(decode_unicode=False, delimiter=b"\0"):
|
for chunk in response.iter_lines(decode_unicode=False, delimiter=b"\0"):
|
||||||
if chunk:
|
if chunk:
|
||||||
data = json.loads(chunk.decode())
|
data = json.loads(chunk.decode())
|
||||||
|
logger.info(f"Response: {data}")
|
||||||
if data["error_code"] == 0:
|
if data["error_code"] == 0:
|
||||||
output = data["text"][skip_echo_len].strip()
|
output = data["text"][skip_echo_len].strip()
|
||||||
output = post_process_code(output)
|
output = post_process_code(output)
|
||||||
|
Loading…
Reference in New Issue
Block a user