fix: can not answer on mac m1-> mps device

This commit is contained in:
yihong0618 2023-05-24 12:33:41 +08:00
parent 0ae64175ef
commit 60ecde5892

View File

@ -9,6 +9,7 @@ from typing import Optional
from pilot.model.compression import compress_module
from pilot.model.adapter import get_llm_model_adapter
from pilot.utils import get_gpu_memory
from pilot.configs.model_config import DEVICE
from pilot.model.llm.monkey_patch import replace_llama_attn_with_non_inplace_operations
def raise_warning_for_incompatible_cpu_offloading_configuration(
@ -50,7 +51,7 @@ class ModelLoader(metaclass=Singleton):
def __init__(self,
model_path) -> None:
self.device = "cuda" if torch.cuda.is_available() else "cpu"
self.device = DEVICE
self.model_path = model_path
self.kwargs = {
"torch_dtype": torch.float16,