mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-08-04 10:00:17 +00:00
fix: proxyllm do not need controller_addr (#616)
reason: for openai user in China, most of us use `http_proxy` or `https_proxy` for openai things, but for the controller_addr load reason when have this proxy in env, the pilot can not start, so we can check if use llmproxy to ignore it
This commit is contained in:
commit
56a3b51fab
@ -1,17 +1,18 @@
|
|||||||
import asyncio
|
import asyncio
|
||||||
import itertools
|
import itertools
|
||||||
import json
|
import json
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import random
|
|
||||||
import time
|
|
||||||
import logging
|
import logging
|
||||||
|
import os
|
||||||
|
import random
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
from concurrent.futures import ThreadPoolExecutor
|
from concurrent.futures import ThreadPoolExecutor
|
||||||
from dataclasses import asdict
|
from dataclasses import asdict
|
||||||
from typing import Awaitable, Callable, Dict, Iterator, List, Optional
|
from typing import Awaitable, Callable, Dict, Iterator, List, Optional
|
||||||
|
|
||||||
from fastapi import APIRouter, FastAPI
|
from fastapi import APIRouter, FastAPI
|
||||||
from fastapi.responses import StreamingResponse
|
from fastapi.responses import StreamingResponse
|
||||||
|
|
||||||
from pilot.component import SystemApp
|
from pilot.component import SystemApp
|
||||||
from pilot.model.base import (
|
from pilot.model.base import (
|
||||||
ModelInstance,
|
ModelInstance,
|
||||||
@ -20,16 +21,16 @@ from pilot.model.base import (
|
|||||||
WorkerApplyType,
|
WorkerApplyType,
|
||||||
WorkerSupportedModel,
|
WorkerSupportedModel,
|
||||||
)
|
)
|
||||||
from pilot.model.cluster.registry import ModelRegistry
|
from pilot.model.cluster.base import *
|
||||||
from pilot.model.llm_utils import list_supported_models
|
|
||||||
from pilot.model.parameter import ModelParameters, ModelWorkerParameters, WorkerType
|
|
||||||
from pilot.model.cluster.worker_base import ModelWorker
|
|
||||||
from pilot.model.cluster.manager_base import (
|
from pilot.model.cluster.manager_base import (
|
||||||
WorkerManager,
|
WorkerManager,
|
||||||
WorkerRunData,
|
|
||||||
WorkerManagerFactory,
|
WorkerManagerFactory,
|
||||||
|
WorkerRunData,
|
||||||
)
|
)
|
||||||
from pilot.model.cluster.base import *
|
from pilot.model.cluster.registry import ModelRegistry
|
||||||
|
from pilot.model.cluster.worker_base import ModelWorker
|
||||||
|
from pilot.model.llm_utils import list_supported_models
|
||||||
|
from pilot.model.parameter import ModelParameters, ModelWorkerParameters, WorkerType
|
||||||
from pilot.utils.parameter_utils import (
|
from pilot.utils.parameter_utils import (
|
||||||
EnvArgumentParser,
|
EnvArgumentParser,
|
||||||
ParameterDescription,
|
ParameterDescription,
|
||||||
@ -639,6 +640,10 @@ def _setup_fastapi(worker_params: ModelWorkerParameters, app=None):
|
|||||||
)
|
)
|
||||||
|
|
||||||
if not worker_params.controller_addr:
|
if not worker_params.controller_addr:
|
||||||
|
# if we have http_proxy or https_proxy in env, the server can not start
|
||||||
|
# so set it to empty here
|
||||||
|
os.environ["http_proxy"] = ""
|
||||||
|
os.environ["https_proxy"] = ""
|
||||||
worker_params.controller_addr = f"http://127.0.0.1:{worker_params.port}"
|
worker_params.controller_addr = f"http://127.0.0.1:{worker_params.port}"
|
||||||
logger.info(
|
logger.info(
|
||||||
f"Run WorkerManager with standalone mode, controller_addr: {worker_params.controller_addr}"
|
f"Run WorkerManager with standalone mode, controller_addr: {worker_params.controller_addr}"
|
||||||
|
@ -1,5 +1,6 @@
|
|||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
import os
|
||||||
from dataclasses import dataclass, field
|
from dataclasses import dataclass, field
|
||||||
from enum import Enum
|
from enum import Enum
|
||||||
from typing import Dict, Optional
|
from typing import Dict, Optional
|
||||||
@ -246,6 +247,11 @@ class ProxyModelParameters(BaseModelParameters):
|
|||||||
proxy_api_key: str = field(
|
proxy_api_key: str = field(
|
||||||
metadata={"tags": "privacy", "help": "The api key of current proxy LLM"},
|
metadata={"tags": "privacy", "help": "The api key of current proxy LLM"},
|
||||||
)
|
)
|
||||||
|
http_proxy: Optional[str] = field(
|
||||||
|
default=os.environ.get("http_proxy") or os.environ.get("https_proxy"),
|
||||||
|
metadata={"help": "The http or https proxy to use openai"},
|
||||||
|
)
|
||||||
|
|
||||||
proxyllm_backend: Optional[str] = field(
|
proxyllm_backend: Optional[str] = field(
|
||||||
default=None,
|
default=None,
|
||||||
metadata={
|
metadata={
|
||||||
|
Loading…
Reference in New Issue
Block a user