Commit Graph

1821 Commits

Author SHA1 Message Date
Jiarui Fang
5682e6d346
[hotfix] correcnt cpu_optim runtime compilation (#2197) 2022-12-26 16:45:14 +08:00
oahzxl
1b8a066592 add chunk select class 2022-12-26 15:28:01 +08:00
BlueRum
6642cebdbe
[example] Change some training settings for diffusion (#2195) 2022-12-26 15:22:20 +08:00
HELSON
2458659919
[zero] fix error for BEiT models (#2169)
* [zero] fix error for BEiT models

* [ColoParameter] add unpack operation for tuple arguments

* fix bugs

* fix chunkv2 unit testing

* add assertion for gradient state
2022-12-26 15:03:54 +08:00
ziyuhuang123
4363ff3e41
'[NFC] fix some typos' (#2175) 2022-12-25 18:41:39 +08:00
binmakeswell
04a200573c
[NFC] update news link (#2191) 2022-12-24 11:53:52 +08:00
Jiarui Fang
355ffb386e
[builder] unified cpu_optim fused_optim inferface (#2190) 2022-12-23 20:57:41 +08:00
oahzxl
786a398a6b code style 2022-12-23 17:42:51 +08:00
oahzxl
9b1b890347 update run 2022-12-23 17:32:11 +08:00
oahzxl
51ef8384c1 finish node reorder 2022-12-23 17:25:36 +08:00
Jiarui Fang
9587b080ba
[builder] use runtime builder for fused_optim (#2189) 2022-12-23 17:07:03 +08:00
oahzxl
884a228ea6 reorder nodes 2022-12-23 17:06:07 +08:00
Fazzie-Maqianli
ce3c4eca7b
[example] support Dreamblooth (#2188) 2022-12-23 16:47:30 +08:00
BlueRum
1cf6d92d7c
[exmaple] diffuser, support quant inference for stable diffusion (#2186) 2022-12-23 16:06:29 +08:00
Jiarui Fang
bc0e271e71
[buider] use builder() for cpu adam and fused optim in setup.py (#2187) 2022-12-23 16:05:13 +08:00
oahzxl
e0ae68e736 code style 2022-12-23 15:49:04 +08:00
oahzxl
fa5e6fbf96 code style 2022-12-23 15:38:37 +08:00
oahzxl
4f5e105af3 remove flow tracer 2022-12-23 15:34:41 +08:00
oahzxl
4d89525fc2 remove abandoned function 2022-12-23 14:28:49 +08:00
oahzxl
49ba619085 code style 2022-12-23 14:26:43 +08:00
oahzxl
d309e9338b adapt codegen to prepose node 2022-12-23 14:26:12 +08:00
Jiarui Fang
d42afd30f8
[builder] runtime adam and fused_optim builder (#2184) 2022-12-23 14:14:21 +08:00
oahzxl
522f017418 code style 2022-12-23 13:41:51 +08:00
oahzxl
774d34f1aa refactor flow search 2022-12-23 13:41:10 +08:00
YuliangLiu0306
550f8f8905
[autoparallel] integrate_gpt_related_tests (#2134)
* [autoparallel] integrate_gpt_related_tests

* polish code

* polish code

* add GPT2Model into runtime test
2022-12-23 12:36:59 +08:00
Ziyue Jiang
59e343328d
[Pipeline Middleware ] Fix deadlock when num_microbatch=num_stage (#2156)
* add splitter

* polish code

* remove comment

* fix async nan by moving to cpu first

Co-authored-by: Ziyue Jiang <ziyue.jiang@gmail.com>
2022-12-23 11:38:43 +08:00
github-actions[bot]
937f404253
Automated submodule synchronization (#2136)
Co-authored-by: github-actions <github-actions@github.com>
2022-12-23 09:34:48 +08:00
Jiarui Fang
65f56f49e8
[example] gpt demo more accuracy tflops (#2178) 2022-12-22 20:51:35 +08:00
Tongping Liu
ab54fed292
[hotfix] add kwargs for colo_addmm (#2171) 2022-12-22 13:25:30 +08:00
Arsmart1
a110933d65
[NFC] fix a typo 'stable-diffusion-typo-fine-tune'
Co-authored-by: ziyuhuang123 <202476410@qq.com>
2022-12-22 10:57:05 +08:00
Fazzie-Maqianli
9396a18361
Merge pull request #2174 from ziyuhuang123/main
'diffusion-typo-change'
2022-12-22 10:37:55 +08:00
ziyuhuang123
cf5028363c 'diffusion-typo-change' 2022-12-22 10:28:59 +08:00
アマデウス
622f863291
[hotfix] Jit type hint #2161 (#2164) 2022-12-22 10:17:03 +08:00
Jiarui Fang
27327a4c90
[example] add palm pytorch version (#2172) 2022-12-22 10:15:34 +08:00
Zihao
12e7bcd720
register meta func for rnn (#2159) 2022-12-21 23:06:18 +08:00
oahzxl
ded1005667 format code 2022-12-21 15:03:08 +08:00
oahzxl
d361d533e8 refactor flow tracer 2022-12-21 15:01:03 +08:00
oahzxl
d734529a39 move flow tracer 2022-12-21 15:00:24 +08:00
Boyuan Yao
cfe2a9bd90
[autoparallel] memory estimation for shape consistency (#2144)
* [fx] metainfo class for auto parallel

* [fx] add unit test for linear metainfo

* [fx] fix bwd param for linear

* [fx] modify unit test

* [fx] modify unit test

* [fx] modify import

* [fx] modify import

* [fx] modify import

* [fx] move meta profiler to auto parallel

* [fx] add conv metainfo class

* [fx] restore profiler

* [fx] restore meta profiler

* [autoparallel] modify unit test

* [fx] modify unit test

* [autoparallel] add batchnorm metainfo class

* [autoparallel] fix batchnorm unit test function declaration

* [fx] restore profiler

* [fx] add relu metainfo class

* [fx] restore profiler

* [autoparallel] modify metainfo input

* [autoparallel] add pooling metainfo

* [autoparallel] add F.linear metainfo generator

* [autoparallel] add binary elementwise metainfo

* [fx] recover profiler

* [autoparallel] fix forward memory calculation

* [autoparallel] modify constants.py

* [autoparallel] remove redundant print

* [autoparallel] add F.conv metainfo

* [autoparallel] linear fix

* [autoparallel] memory estimation for communication actions

* [autoparallel] fix docstring

* [autoparallel] fix variables name
2022-12-21 10:39:37 +08:00
Jiarui Fang
b87496a66b
[hotfix] fix auto policy of test_sharded_optim_v2 (#2157) 2022-12-20 23:03:18 +08:00
YuliangLiu0306
16335cb537
[hotfix] fix aten default bug (#2158) 2022-12-20 22:40:46 +08:00
Jiarui Fang
a4b4bb01d6
[example] update vit readme (#2155) 2022-12-20 15:56:54 +08:00
Jiarui Fang
2cfe685b9f
[exmaple] add vit missing functions (#2154) 2022-12-20 15:03:26 +08:00
HELSON
a7d95b7024
[example] add zero1, zero2 example in GPT examples (#2146)
* [example] add zero1 and zero2 for GPT

* update readme in gpt example

* polish code

* change init value

* update readme
2022-12-20 14:30:27 +08:00
YuliangLiu0306
1cce6e36ca
[autoparallel] use metainfo in handler (#2149) 2022-12-20 10:31:22 +08:00
Jiarui Fang
9b39170a5c
[version] 0.1.13 (#2152) 2022-12-20 10:28:04 +08:00
Jiarui Fang
e0c01d1db1
Revert "[version] version to v0.1.13 (#2139)" (#2153)
This reverts commit 6ad866b684.
2022-12-20 10:26:36 +08:00
Jiarui Fang
2827f41898
[Gemini] GeminiDPP convert to PyTorch Module. (#2151) 2022-12-20 10:19:36 +08:00
Jiarui Fang
bdef9dfdbe
[NFC] remove useless graph node code (#2150) 2022-12-20 00:33:58 +08:00
BlueRum
b3f73ce1c8
[Gemini] Update coloinit_ctx to support meta_tensor (#2147) 2022-12-19 22:37:07 +08:00