From e239cf9060d90a22557d685c9d6ce350520c5337 Mon Sep 17 00:00:00 2001 From: Luo Yihang Date: Tue, 5 Mar 2024 21:44:38 +0800 Subject: [PATCH] [hotfix] fix typo of openmoe model source (#5403) --- examples/language/openmoe/benchmark/benchmark_cai.py | 2 +- examples/language/openmoe/benchmark/benchmark_fsdp.py | 2 +- examples/language/openmoe/infer.py | 6 +++--- examples/language/openmoe/train.py | 4 ++-- 4 files changed, 7 insertions(+), 7 deletions(-) diff --git a/examples/language/openmoe/benchmark/benchmark_cai.py b/examples/language/openmoe/benchmark/benchmark_cai.py index 03b660ecf..770c500d8 100644 --- a/examples/language/openmoe/benchmark/benchmark_cai.py +++ b/examples/language/openmoe/benchmark/benchmark_cai.py @@ -207,7 +207,7 @@ def main(): coordinator.print_on_master(f"Set plugin as {plugin}") # Build OpenMoe model - repo_name = "hpcaitech/openmoe-" + args.model_name + repo_name = "hpcai-tech/openmoe-" + args.model_name config = LlamaConfig.from_pretrained(repo_name) set_openmoe_args( config, diff --git a/examples/language/openmoe/benchmark/benchmark_fsdp.py b/examples/language/openmoe/benchmark/benchmark_fsdp.py index 7f438fc5a..b00fbd001 100644 --- a/examples/language/openmoe/benchmark/benchmark_fsdp.py +++ b/examples/language/openmoe/benchmark/benchmark_fsdp.py @@ -53,7 +53,7 @@ def fsdp_main(rank, world_size, args): train_loader = torch.utils.data.DataLoader(dataset, **train_kwargs) torch.cuda.set_device(rank) - config = LlamaConfig.from_pretrained("hpcaitech/openmoe-%s" % args.model_name) + config = LlamaConfig.from_pretrained("hpcai-tech/openmoe-%s" % args.model_name) set_openmoe_args( config, num_experts=config.num_experts, diff --git a/examples/language/openmoe/infer.py b/examples/language/openmoe/infer.py index db90c6e34..04df64531 100644 --- a/examples/language/openmoe/infer.py +++ b/examples/language/openmoe/infer.py @@ -15,19 +15,19 @@ def parse_args(): def inference(args): tokenizer = T5Tokenizer.from_pretrained("google/umt5-small") if args.model == "test": - config = LlamaConfig.from_pretrained("hpcaitech/openmoe-base") + config = LlamaConfig.from_pretrained("hpcai-tech/openmoe-base") set_openmoe_args(config, num_experts=config.num_experts, moe_layer_interval=config.moe_layer_interval, enable_kernel=True) model = OpenMoeForCausalLM(config) else: - config = LlamaConfig.from_pretrained(f"hpcaitech/openmoe-{args.model}") + config = LlamaConfig.from_pretrained(f"hpcai-tech/openmoe-{args.model}") set_openmoe_args(config, num_experts=config.num_experts, moe_layer_interval=config.moe_layer_interval, enable_kernel=False) - model = OpenMoeForCausalLM.from_pretrained(f"hpcaitech/openmoe-{args.model}", config=config) + model = OpenMoeForCausalLM.from_pretrained(f"hpcai-tech/openmoe-{args.model}", config=config) model = model.eval().bfloat16() model = model.to(torch.cuda.current_device()) diff --git a/examples/language/openmoe/train.py b/examples/language/openmoe/train.py index 1ae661f54..89c4d5420 100644 --- a/examples/language/openmoe/train.py +++ b/examples/language/openmoe/train.py @@ -269,12 +269,12 @@ def main(): # Build OpenMoe model if test_mode: - config = LlamaConfig.from_pretrained("hpcaitech/openmoe-base") + config = LlamaConfig.from_pretrained("hpcai-tech/openmoe-base") config.hidden_size = 128 config.intermediate_size = 256 config.vocab_size = 32000 else: - repo_name = "hpcaitech/openmoe-" + args.model_name + repo_name = "hpcai-tech/openmoe-" + args.model_name config = LlamaConfig.from_pretrained(repo_name) set_openmoe_args( config,