Yuanheng Zhao
5d4c1fe8f5
[Fix/Inference] Fix GQA Triton and Support Llama3 (#5624)
* [fix] GQA calling of flash decoding triton
* fix kv cache alloc shape
* fix rotary triton - GQA
* fix sequence max length assigning
* Sequence max length logic
* fix scheduling and spec-dec
* skip without import error
* fix pytest - skip without ImportError
---------
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-04-23 13:09:55 +08:00
..
2024-04-08 15:09:40 +08:00
2023-09-19 14:20:26 +08:00
2024-01-09 10:20:05 +08:00
2023-09-19 14:20:26 +08:00
2024-03-27 13:57:00 +08:00
2024-04-08 15:09:40 +08:00
2024-04-03 17:15:47 +08:00
2023-09-19 14:20:26 +08:00
2023-09-19 14:20:26 +08:00
2023-09-19 14:20:26 +08:00
2024-04-08 15:09:40 +08:00
2024-04-23 13:09:55 +08:00
2024-04-08 15:09:40 +08:00
2024-01-09 10:20:05 +08:00
2024-03-25 12:31:09 +08:00
2024-04-08 15:09:40 +08:00
2024-04-08 15:09:40 +08:00
2024-04-03 17:15:47 +08:00
2023-10-16 11:28:44 +08:00
2024-03-26 17:22:27 +08:00
2024-01-09 10:20:05 +08:00
2022-03-11 15:50:28 +08:00