mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2026-05-04 01:48:43 +00:00
* [inference] Dynamic Batching for Single and Multiple GPUs (#4831) * finish batch manager * 1 * first * fix * fix dynamic batching * llama infer * finish test * support different lengths generating * del prints * del prints * fix * fix bug --------- Co-authored-by: CjhHa1 <cjh18671720497outlook.com> * [inference] Async dynamic batching (#4894) * finish input and output logic * add generate * test forward * 1 * [inference]Re push async dynamic batching (#4901) * adapt to ray server * finish async * finish test * del test --------- Co-authored-by: yuehuayingxueluo <867460659@qq.com> * Revert "[inference]Re push async dynamic batching (#4901)" (#4905) This reverts commitfbf3c09e67. * Revert "[inference] Async dynamic batching (#4894)" This reverts commitfced140250. * Revert "[inference] Async dynamic batching (#4894)" (#4909) This reverts commitfced140250. * Add Ray Distributed Environment Init Scripts * support DynamicBatchManager base function * revert _set_tokenizer version * add driver async generate * add async test * fix bugs in test_ray_dist.py * add get_tokenizer.py * fix code style * fix bugs about No module named 'pydantic' in ci test * fix bugs in ci test * fix bugs in ci test * fix bugs in ci test * [infer]Add Ray Distributed Environment Init Scripts (#4911) * Revert "[inference] Async dynamic batching (#4894)" This reverts commitfced140250. * Add Ray Distributed Environment Init Scripts * support DynamicBatchManager base function * revert _set_tokenizer version * add driver async generate * add async test * fix bugs in test_ray_dist.py * add get_tokenizer.py * fix code style * fix bugs about No module named 'pydantic' in ci test * fix bugs in ci test * fix bugs in ci test * fix bugs in ci test * support dynamic batch for bloom model and is_running function * [Inference]Test for new Async engine (#4935) * infer engine * infer engine * test engine * test engine * new manager * change step * add * test * fix * fix * finish test * finish test * finish test * finish test * add license --------- Co-authored-by: yuehuayingxueluo <867460659@qq.com> * add assertion for config (#4947) * [Inference] Finish dynamic batching offline test (#4948) * test * fix test * fix quant * add default * fix * fix some bugs * fix some bugs * fix * fix bug * fix bugs * reset param --------- Co-authored-by: yuehuayingxueluo <867460659@qq.com> Co-authored-by: Cuiqing Li <lixx3527@gmail.com> Co-authored-by: CjhHa1 <cjh18671720497outlook.com>
62 lines
1.9 KiB
Python
62 lines
1.9 KiB
Python
import asyncio
|
|
import os
|
|
import uuid
|
|
|
|
import pytest
|
|
|
|
import colossalai
|
|
from colossalai.inference.async_engine import Async_Engine
|
|
from colossalai.inference.dynamic_batching.ray_init_config import RayInitConfig
|
|
from colossalai.inference.dynamic_batching.sampling_params import SamplingParams
|
|
from colossalai.testing import clear_cache_before_run, rerun_if_address_is_in_use, spawn
|
|
|
|
PATH = "config.yaml"
|
|
|
|
|
|
def run_async_engine(path: str):
|
|
if not os.path.exists(path):
|
|
return
|
|
|
|
config = RayInitConfig.from_yaml_path(path)
|
|
engine_config = config.engine_config_data
|
|
model = engine_config.model
|
|
if model is None or not os.path.exists(model):
|
|
return
|
|
|
|
prompt = "Introduce some landmarks in London.\n The Tower of London is a historic castle on the north bank of the River Thames in central London. It was founded towards the end of 10"
|
|
sampling_params = SamplingParams()
|
|
asyncio.run(asy_for_loop_test(config, prompt, sampling_params))
|
|
|
|
|
|
async def get_result(engine, prompt, sampling_params):
|
|
request_id = str(uuid.uuid4().hex)
|
|
results = engine.generate(request_id, prompt, sampling_params)
|
|
async for result in results:
|
|
# print(result)
|
|
assert result is not None
|
|
|
|
|
|
async def asy_for_loop_test(config, prompt, sampling_params):
|
|
router_config = config.router_config_data
|
|
engine_config = config.engine_config_data
|
|
engine = Async_Engine(router_config=router_config, engine_config=engine_config)
|
|
for i in range(10):
|
|
print("in for loop", i)
|
|
await get_result(engine, prompt, sampling_params)
|
|
|
|
|
|
def check_async_engine(rank, world_size, port):
|
|
colossalai.launch(config={}, rank=rank, world_size=world_size, host="localhost", port=port, backend="nccl")
|
|
run_async_engine(PATH)
|
|
|
|
|
|
@pytest.mark.dist
|
|
@rerun_if_address_is_in_use()
|
|
@clear_cache_before_run()
|
|
def test_async_engine():
|
|
spawn(check_async_engine, 1)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
test_async_engine()
|