feat(model): Support Qwen2MoE (#1439)

Co-authored-by: Fangyin Cheng <staneyffer@gmail.com>
This commit is contained in:
yyhhyy
2024-04-20 21:03:10 +08:00
committed by GitHub
parent 82e4ce4c43
commit dce03862d5
4 changed files with 32 additions and 0 deletions

View File

@@ -114,6 +114,8 @@ LLM_MODEL_CONFIG = {
"qwen1.5-72b-chat": os.path.join(MODEL_PATH, "Qwen1.5-72B-Chat"),
# https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat
"codeqwen1.5-7b-chat": os.path.join(MODEL_PATH, "CodeQwen1.5-7B-Chat"),
# https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B-Chat
"qwen1.5-moe-a2.7b-chat": os.path.join(MODEL_PATH, "Qwen1.5-MoE-A2.7B-Chat"),
# (Llama2 based) We only support WizardLM-13B-V1.2 for now, which is trained from Llama-2 13b, see https://huggingface.co/WizardLM/WizardLM-13B-V1.2
"wizardlm-13b": os.path.join(MODEL_PATH, "WizardLM-13B-V1.2"),
# wget https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF/resolve/main/vicuna-13b-v1.5.Q4_K_M.gguf -O models/ggml-model-q4_0.gguf