diff --git a/pilot/model/cluster/worker/manager.py b/pilot/model/cluster/worker/manager.py index 2f84a3216..da34c314a 100644 --- a/pilot/model/cluster/worker/manager.py +++ b/pilot/model/cluster/worker/manager.py @@ -639,7 +639,11 @@ def _setup_fastapi(worker_params: ModelWorkerParameters, app=None): router as controller_router, ) - if not worker_params.controller_addr and worker_params.model_name != "proxyllm": + if not worker_params.controller_addr: + # if we have http_proxy or https_proxy in env, the server can not start + # so set it to empty here + os.environ["http_proxy"] = "" + os.environ["https_proxy"] = "" worker_params.controller_addr = f"http://127.0.0.1:{worker_params.port}" logger.info( f"Run WorkerManager with standalone mode, controller_addr: {worker_params.controller_addr}" diff --git a/pilot/model/parameter.py b/pilot/model/parameter.py index c7917a95f..6fd8ec5ab 100644 --- a/pilot/model/parameter.py +++ b/pilot/model/parameter.py @@ -1,5 +1,6 @@ #!/usr/bin/env python3 # -*- coding: utf-8 -*- +import os from dataclasses import dataclass, field from enum import Enum from typing import Dict, Optional @@ -246,6 +247,11 @@ class ProxyModelParameters(BaseModelParameters): proxy_api_key: str = field( metadata={"tags": "privacy", "help": "The api key of current proxy LLM"}, ) + http_proxy: Optional[str] = field( + default=os.environ.get("http_proxy") or os.environ.get("https_proxy"), + metadata={"help": "The http or https proxy to use openai"}, + ) + proxyllm_backend: Optional[str] = field( default=None, metadata={