1
0
mirror of https://github.com/hpcaitech/ColossalAI.git synced 2025-05-07 16:08:22 +00:00
Commit Graph

21 Commits

Author SHA1 Message Date
Hongxin Liu
079bf3cb26
[misc] update pre-commit and run all files ()
* [misc] update pre-commit

* [misc] run pre-commit

* [misc] remove useless configuration files

* [misc] ignore cuda for clang-format
2023-09-19 14:20:26 +08:00
Hongxin Liu
b5f9e37c70
[legacy] clean up legacy code ()
* [legacy] remove outdated codes of pipeline ()

* [legacy] remove cli of benchmark and update optim ()

* [legacy] remove cli of benchmark and update optim

* [doc] fix cli doc test

* [legacy] fix engine clip grad norm

* [legacy] remove outdated colo tensor ()

* [legacy] remove outdated colo tensor

* [test] fix test import

* [legacy] move outdated zero to legacy ()

* [legacy] clean up utils ()

* [legacy] clean up utils

* [example] update examples

* [legacy] clean up amp

* [legacy] fix amp module

* [legacy] clean up gpc ()

* [legacy] clean up context

* [legacy] clean core, constants and global vars

* [legacy] refactor initialize

* [example] fix examples ci

* [example] fix examples ci

* [legacy] fix tests

* [example] fix gpt example

* [example] fix examples ci

* [devops] fix ci installation

* [example] fix examples ci
2023-09-18 16:31:06 +08:00
digger-yu
b9a8dff7e5
[doc] Fix typo under colossalai and doc()
* Fixed several spelling errors under colossalai

* Fix the spelling error in colossalai and docs directory

* Cautious Changed the spelling error under the example folder

* Update runtime_preparation_pass.py

revert autograft to autograd

* Update search_chunk.py

utile to until

* Update check_installation.py

change misteach to mismatch in line 91

* Update 1D_tensor_parallel.md

revert to perceptron

* Update 2D_tensor_parallel.md

revert to perceptron in line 73

* Update 2p5D_tensor_parallel.md

revert to perceptron in line 71

* Update 3D_tensor_parallel.md

revert to perceptron in line 80

* Update README.md

revert to resnet in line 42

* Update reorder_graph.py

revert to indice in line 7

* Update p2p.py

revert to megatron in line 94

* Update initialize.py

revert to torchrun in line 198

* Update routers.py

change to detailed in line 63

* Update routers.py

change to detailed in line 146

* Update README.md

revert  random number in line 402
2023-04-26 11:38:43 +08:00
ver217
26b7aac0be
[zero] reorganize zero/gemini folder structure ()
* [zero] refactor low-level zero folder structure

* [zero] fix legacy zero import path

* [zero] fix legacy zero import path

* [zero] remove useless import

* [zero] refactor gemini folder structure

* [zero] refactor gemini folder structure

* [zero] refactor legacy zero import path

* [zero] refactor gemini folder structure

* [zero] refactor gemini folder structure

* [zero] refactor gemini folder structure

* [zero] refactor legacy zero import path

* [zero] fix test import path

* [zero] fix test

* [zero] fix circular import

* [zero] update import
2023-04-04 13:48:16 +08:00
HELSON
1a1d68b053
[moe] add checkpoint for moe models ()
* [moe] add checkpoint for moe models

* [hotfix] fix bugs in unit test
2023-03-31 09:20:33 +08:00
HELSON
a088022efc
[moe] fix moe bugs () 2022-09-23 15:33:57 +08:00
HELSON
f7f2248771
[moe] fix MoE bugs ()
* remove forced FP32 modules

* correct no_shard-contexts' positions
2022-09-22 13:56:30 +08:00
ver217
8432dc7080
polish moe docsrting () 2022-04-01 16:15:36 +08:00
HELSON
e6d50ec107
[zero] adapt zero for unsharded parameters ()
* support existing sharded and unsharded parameters in zero

* add unitest for moe-zero model init

* polish moe gradient handler
2022-03-31 18:34:11 +08:00
HELSON
8c90d4df54
[zero] add zero context manager to change config during initialization () 2022-03-29 17:57:59 +08:00
Liang Bowen
ec5086c49c Refactored docstring to google style 2022-03-29 17:17:47 +08:00
Jiarui Fang
a445e118cf
[polish] polish singleton and global context () 2022-03-23 18:03:39 +08:00
HELSON
c9023d4078
[MOE] support PR-MOE () 2022-03-22 16:48:22 +08:00
HELSON
d7ea63992b
[MOE] add FP32LinearGate for MOE in NaiveAMP context () 2022-03-22 10:50:20 +08:00
HELSON
aff9d354f7
[MOE] polish moe_env () 2022-03-19 15:36:25 +08:00
HELSON
dbdc9a7783
added Multiply Jitter and capacity factor eval for MOE () 2022-03-16 16:47:44 +08:00
HELSON
3f70a2b12f
removed noisy function during evaluation of MoE router () 2022-03-15 12:06:09 +08:00
1SAA
82023779bb Added TPExpert for special situation 2022-03-11 15:50:28 +08:00
1SAA
219df6e685 Optimized MoE layer and fixed some bugs;
Decreased moe tests;

Added FFNExperts and ViTMoE model
2022-03-11 15:50:28 +08:00
HELSON
0f8c7f9804
Fixed docstring in colossalai () 2022-01-21 10:44:30 +08:00
HELSON
dceae85195
Added MoE parallel () 2022-01-07 15:08:36 +08:00