feat(model): Support CodeQwen1.5-7B-Chat (#1431)

This commit is contained in:
yyhhyy
2024-04-18 15:14:14 +08:00
committed by GitHub
parent 461607e421
commit 8625690107
3 changed files with 4 additions and 0 deletions

View File

@@ -108,6 +108,8 @@ LLM_MODEL_CONFIG = {
# https://huggingface.co/Qwen/Qwen1.5-32B-Chat
"qwen1.5-32b-chat": os.path.join(MODEL_PATH, "Qwen1.5-32B-Chat"),
"qwen1.5-72b-chat": os.path.join(MODEL_PATH, "Qwen1.5-72B-Chat"),
# https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat
"codeqwen1.5-7b-chat": os.path.join(MODEL_PATH, "CodeQwen1.5-7B-Chat"),
# (Llama2 based) We only support WizardLM-13B-V1.2 for now, which is trained from Llama-2 13b, see https://huggingface.co/WizardLM/WizardLM-13B-V1.2
"wizardlm-13b": os.path.join(MODEL_PATH, "WizardLM-13B-V1.2"),
# wget https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF/resolve/main/vicuna-13b-v1.5.Q4_K_M.gguf -O models/ggml-model-q4_0.gguf