ColossalAI/tests/test_infer/test_kernels/cuda
Li Xingjian 8554585a5f
[Inference] Fix flash-attn import and add model test (#5794)
* Fix torch int32 dtype

Signed-off-by: char-1ee <xingjianli59@gmail.com>

* Fix flash-attn import

Signed-off-by: char-1ee <xingjianli59@gmail.com>

* Add generalized model test

Signed-off-by: char-1ee <xingjianli59@gmail.com>

* Remove exposed path to model

Signed-off-by: char-1ee <xingjianli59@gmail.com>

* Add default value for use_flash_attn

Signed-off-by: char-1ee <xingjianli59@gmail.com>

* Rename model test

Signed-off-by: char-1ee <xingjianli59@gmail.com>

---------

Signed-off-by: char-1ee <xingjianli59@gmail.com>
2024-06-12 14:13:50 +08:00
..
__init__.py [Fix] Fix & Update Inference Tests (compatibility w/ main) 2024-05-05 16:28:56 +00:00
test_convert_fp8.py [Inference/Feat] Add convert_fp8 op for fp8 test in the future (#5706) 2024-05-10 18:39:54 +08:00
test_flash_decoding_attention.py Fix test import 2024-06-10 02:03:30 +00:00
test_get_cos_and_sin.py [Fix] Fix & Update Inference Tests (compatibility w/ main) 2024-05-05 16:28:56 +00:00
test_kv_cache_memcpy.py [Inference] Fix flash-attn import and add model test (#5794) 2024-06-12 14:13:50 +08:00
test_rms_layernorm.py [Fix] Fix & Update Inference Tests (compatibility w/ main) 2024-05-05 16:28:56 +00:00
test_rotary_embdding_unpad.py [Fix] Fix & Update Inference Tests (compatibility w/ main) 2024-05-05 16:28:56 +00:00
test_silu_and_mul.py [Fix] Fix & Update Inference Tests (compatibility w/ main) 2024-05-05 16:28:56 +00:00