feat: Check serialization for AWEL operator function

This commit is contained in:
Fangyin Cheng 2024-09-04 22:08:55 +08:00
parent b1ca247419
commit f8ce7d4580
12 changed files with 236 additions and 4 deletions

View File

@ -1,3 +1,6 @@
.env
.git/
./.mypy_cache/
models/ models/
plugins/ plugins/
pilot/data pilot/data
@ -5,6 +8,8 @@ pilot/message
logs/ logs/
venv/ venv/
web/node_modules/ web/node_modules/
web/.next/
web/.env
docs/node_modules/ docs/node_modules/
build/ build/
docs/build/ docs/build/

View File

@ -332,6 +332,10 @@ class Config(metaclass=Singleton):
os.getenv("MULTI_INSTANCE", "False").lower() == "true" os.getenv("MULTI_INSTANCE", "False").lower() == "true"
) )
self.SCHEDULER_ENABLED = (
os.getenv("SCHEDULER_ENABLED", "True").lower() == "true"
)
@property @property
def local_db_manager(self) -> "ConnectorManager": def local_db_manager(self) -> "ConnectorManager":
from dbgpt.datasource.manages import ConnectorManager from dbgpt.datasource.manages import ConnectorManager

View File

@ -19,12 +19,14 @@ class DefaultScheduler(BaseComponent):
system_app: SystemApp, system_app: SystemApp,
scheduler_delay_ms: int = 5000, scheduler_delay_ms: int = 5000,
scheduler_interval_ms: int = 1000, scheduler_interval_ms: int = 1000,
scheduler_enable: bool = True,
): ):
super().__init__(system_app) super().__init__(system_app)
self.system_app = system_app self.system_app = system_app
self._scheduler_interval_ms = scheduler_interval_ms self._scheduler_interval_ms = scheduler_interval_ms
self._scheduler_delay_ms = scheduler_delay_ms self._scheduler_delay_ms = scheduler_delay_ms
self._stop_event = threading.Event() self._stop_event = threading.Event()
self._scheduler_enable = scheduler_enable
def init_app(self, system_app: SystemApp): def init_app(self, system_app: SystemApp):
self.system_app = system_app self.system_app = system_app
@ -39,7 +41,7 @@ class DefaultScheduler(BaseComponent):
def _scheduler(self): def _scheduler(self):
time.sleep(self._scheduler_delay_ms / 1000) time.sleep(self._scheduler_delay_ms / 1000)
while not self._stop_event.is_set(): while self._scheduler_enable and not self._stop_event.is_set():
try: try:
schedule.run_pending() schedule.run_pending()
except Exception as e: except Exception as e:

View File

@ -145,6 +145,9 @@ class DAGVar:
_executor: Optional[Executor] = None _executor: Optional[Executor] = None
_variables_provider: Optional["VariablesProvider"] = None _variables_provider: Optional["VariablesProvider"] = None
# Whether check serializable for AWEL, it will be set to True when running AWEL
# operator in remote environment
_check_serializable: Optional[bool] = None
@classmethod @classmethod
def enter_dag(cls, dag) -> None: def enter_dag(cls, dag) -> None:
@ -257,6 +260,24 @@ class DAGVar:
""" """
cls._variables_provider = variables_provider cls._variables_provider = variables_provider
@classmethod
def get_check_serializable(cls) -> Optional[bool]:
"""Get the check serializable flag.
Returns:
Optional[bool]: The check serializable flag
"""
return cls._check_serializable
@classmethod
def set_check_serializable(cls, check_serializable: bool) -> None:
"""Set the check serializable flag.
Args:
check_serializable (bool): The check serializable flag to set
"""
cls._check_serializable = check_serializable
class DAGLifecycle: class DAGLifecycle:
"""The lifecycle of DAG.""" """The lifecycle of DAG."""
@ -286,6 +307,7 @@ class DAGNode(DAGLifecycle, DependencyMixin, ViewMixin, ABC):
node_name: Optional[str] = None, node_name: Optional[str] = None,
system_app: Optional[SystemApp] = None, system_app: Optional[SystemApp] = None,
executor: Optional[Executor] = None, executor: Optional[Executor] = None,
check_serializable: Optional[bool] = None,
**kwargs, **kwargs,
) -> None: ) -> None:
"""Initialize a DAGNode. """Initialize a DAGNode.
@ -311,6 +333,7 @@ class DAGNode(DAGLifecycle, DependencyMixin, ViewMixin, ABC):
node_id = self._dag._new_node_id() node_id = self._dag._new_node_id()
self._node_id: Optional[str] = node_id self._node_id: Optional[str] = node_id
self._node_name: Optional[str] = node_name self._node_name: Optional[str] = node_name
self._check_serializable = check_serializable
if self._dag: if self._dag:
self._dag._append_node(self) self._dag._append_node(self)
@ -486,6 +509,20 @@ class DAGNode(DAGLifecycle, DependencyMixin, ViewMixin, ABC):
"""Return the string of current DAGNode.""" """Return the string of current DAGNode."""
return self.__repr__() return self.__repr__()
@classmethod
def _do_check_serializable(cls, obj: Any, obj_name: str = "Object"):
"""Check whether the current DAGNode is serializable."""
from dbgpt.util.serialization.check import check_serializable
check_serializable(obj, obj_name)
@property
def check_serializable(self) -> bool:
"""Whether check serializable for current DAGNode."""
if self._check_serializable is not None:
return self._check_serializable or False
return DAGVar.get_check_serializable() or False
def _build_task_key(task_name: str, key: str) -> str: def _build_task_key(task_name: str, key: str) -> str:
return f"{task_name}___$$$$$$___{key}" return f"{task_name}___$$$$$$___{key}"

View File

@ -193,12 +193,29 @@ class BaseOperator(DAGNode, ABC, Generic[OUT], metaclass=BaseOperatorMeta):
self.incremental_output = bool(kwargs["incremental_output"]) self.incremental_output = bool(kwargs["incremental_output"])
if "output_format" in kwargs: if "output_format" in kwargs:
self.output_format = kwargs["output_format"] self.output_format = kwargs["output_format"]
self._runner: WorkflowRunner = runner self._runner: WorkflowRunner = runner
self._dag_ctx: Optional[DAGContext] = None self._dag_ctx: Optional[DAGContext] = None
self._can_skip_in_branch = can_skip_in_branch self._can_skip_in_branch = can_skip_in_branch
self._variables_provider = variables_provider self._variables_provider = variables_provider
def __getstate__(self):
"""Customize the pickling process."""
state = self.__dict__.copy()
if "_runner" in state:
del state["_runner"]
if "_executor" in state:
del state["_executor"]
if "_system_app" in state:
del state["_system_app"]
return state
def __setstate__(self, state):
"""Customize the unpickling process."""
self.__dict__.update(state)
self._runner = default_runner
self._system_app = DAGVar.get_current_system_app()
self._executor = DAGVar.get_executor()
@property @property
def current_dag_context(self) -> DAGContext: def current_dag_context(self) -> DAGContext:
"""Return the current DAG context.""" """Return the current DAG context."""

View File

@ -41,6 +41,12 @@ class JoinOperator(BaseOperator, Generic[OUT]):
super().__init__(can_skip_in_branch=can_skip_in_branch, **kwargs) super().__init__(can_skip_in_branch=can_skip_in_branch, **kwargs)
if not callable(combine_function): if not callable(combine_function):
raise ValueError("combine_function must be callable") raise ValueError("combine_function must be callable")
if self.check_serializable:
super()._do_check_serializable(
combine_function,
f"JoinOperator: {self}, combine_function: {combine_function}",
)
self.combine_function = combine_function self.combine_function = combine_function
async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]: async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]:
@ -83,6 +89,11 @@ class ReduceStreamOperator(BaseOperator, Generic[IN, OUT]):
super().__init__(**kwargs) super().__init__(**kwargs)
if reduce_function and not callable(reduce_function): if reduce_function and not callable(reduce_function):
raise ValueError("reduce_function must be callable") raise ValueError("reduce_function must be callable")
if reduce_function and self.check_serializable:
super()._do_check_serializable(
reduce_function, f"Operator: {self}, reduce_function: {reduce_function}"
)
self.reduce_function = reduce_function self.reduce_function = reduce_function
async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]: async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]:
@ -133,6 +144,12 @@ class MapOperator(BaseOperator, Generic[IN, OUT]):
super().__init__(**kwargs) super().__init__(**kwargs)
if map_function and not callable(map_function): if map_function and not callable(map_function):
raise ValueError("map_function must be callable") raise ValueError("map_function must be callable")
if map_function and self.check_serializable:
super()._do_check_serializable(
map_function, f"Operator: {self}, map_function: {map_function}"
)
self.map_function = map_function self.map_function = map_function
async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]: async def _do_run(self, dag_ctx: DAGContext) -> TaskOutput[OUT]:

View File

@ -94,6 +94,17 @@ class ProxyLLMClient(LLMClient):
self.executor = executor or ThreadPoolExecutor() self.executor = executor or ThreadPoolExecutor()
self.proxy_tokenizer = proxy_tokenizer or TiktokenProxyTokenizer() self.proxy_tokenizer = proxy_tokenizer or TiktokenProxyTokenizer()
def __getstate__(self):
"""Customize the serialization of the object"""
state = self.__dict__.copy()
state.pop("executor")
return state
def __setstate__(self, state):
"""Customize the deserialization of the object"""
self.__dict__.update(state)
self.executor = ThreadPoolExecutor()
@classmethod @classmethod
@abstractmethod @abstractmethod
def new_client( def new_client(

View File

@ -341,7 +341,7 @@ class BuiltinAgentsVariablesProvider(BuiltinVariablesProvider):
StorageVariables( StorageVariables(
key=key, key=key,
name=agent["name"], name=agent["name"],
label=agent["desc"], label=agent["name"],
value=agent["name"], value=agent["name"],
scope=scope, scope=scope,
scope_key=scope_key, scope_key=scope_key,

View File

@ -285,6 +285,9 @@ class BaseDao(Generic[T, REQ, RES]):
else model_to_dict(query_request) else model_to_dict(query_request)
) )
for key, value in query_dict.items(): for key, value in query_dict.items():
if value and isinstance(value, (list, tuple, dict, set)):
# Skip the list, tuple, dict, set
continue
if value is not None and hasattr(model_cls, key): if value is not None and hasattr(model_cls, key):
if isinstance(value, list): if isinstance(value, list):
if len(value) > 0: if len(value) > 0:

View File

@ -1,5 +1,6 @@
import errno import errno
import socket import socket
from typing import Set, Tuple
def _get_ip_address(address: str = "10.254.254.254:1") -> str: def _get_ip_address(address: str = "10.254.254.254:1") -> str:
@ -22,3 +23,34 @@ def _get_ip_address(address: str = "10.254.254.254:1") -> str:
finally: finally:
s.close() s.close()
return curr_address return curr_address
async def _async_get_free_port(
port_range: Tuple[int, int], timeout: int, used_ports: Set[int]
):
import asyncio
loop = asyncio.get_running_loop()
return await loop.run_in_executor(
None, _get_free_port, port_range, timeout, used_ports
)
def _get_free_port(port_range: Tuple[int, int], timeout: int, used_ports: Set[int]):
import random
available_ports = set(range(port_range[0], port_range[1] + 1)) - used_ports
if not available_ports:
raise RuntimeError("No available ports in the specified range")
while available_ports:
port = random.choice(list(available_ports))
try:
with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:
s.bind(("", port))
used_ports.add(port)
return port
except OSError:
available_ports.remove(port)
raise RuntimeError("No available ports in the specified range")

View File

@ -0,0 +1,85 @@
import inspect
from io import StringIO
from typing import Any, Dict, Optional, TextIO
import cloudpickle
def check_serializable(
obj: Any, obj_name: str = "Object", error_msg: str = "Object is not serializable"
):
try:
cloudpickle.dumps(obj)
except Exception as e:
inspect_info = inspect_serializability(obj, obj_name)
msg = f"{error_msg}\n{inspect_info['report']}"
raise TypeError(msg) from e
class SerializabilityInspector:
def __init__(self, stream: Optional[TextIO] = None):
self.stream = stream or StringIO()
self.failures = {}
self.indent_level = 0
def log(self, message: str):
indent = " " * self.indent_level
self.stream.write(f"{indent}{message}\n")
def inspect(self, obj: Any, name: str, depth: int = 3) -> bool:
self.log(f"Inspecting '{name}'")
self.indent_level += 1
try:
cloudpickle.dumps(obj)
self.indent_level -= 1
return True
except Exception as e:
self.failures[name] = str(e)
self.log(f"Failure: {str(e)}")
if depth > 0:
if inspect.isfunction(obj) or inspect.ismethod(obj):
self._inspect_function(obj, depth - 1)
elif hasattr(obj, "__dict__"):
self._inspect_object(obj, depth - 1)
self.indent_level -= 1
return False
def _inspect_function(self, func, depth):
closure = inspect.getclosurevars(func)
for name, value in closure.nonlocals.items():
self.inspect(value, f"{func.__name__}.{name}", depth)
for name, value in closure.globals.items():
self.inspect(value, f"global:{name}", depth)
def _inspect_object(self, obj, depth):
for name, value in inspect.getmembers(obj):
if not name.startswith("__"):
self.inspect(value, f"{type(obj).__name__}.{name}", depth)
def get_report(self) -> str:
summary = "\nSummary of Serialization Failures:\n"
if not self.failures:
summary += "All components are serializable.\n"
else:
for name, error in self.failures.items():
summary += f" - {name}: {error}\n"
return self.stream.getvalue() + summary
def inspect_serializability(
obj: Any,
name: Optional[str] = None,
depth: int = 5,
stream: Optional[TextIO] = None,
) -> Dict[str, Any]:
inspector = SerializabilityInspector(stream)
success = inspector.inspect(obj, name or type(obj).__name__, depth)
return {
"success": success,
"failures": inspector.failures,
"report": inspector.get_report(),
}

View File

@ -20,16 +20,21 @@ LOAD_EXAMPLES="true"
BUILD_NETWORK="" BUILD_NETWORK=""
DB_GPT_INSTALL_MODEL="default" DB_GPT_INSTALL_MODEL="default"
DOCKERFILE="Dockerfile"
IMAGE_NAME_SUFFIX=""
usage () { usage () {
echo "USAGE: $0 [--base-image nvidia/cuda:12.1.0-runtime-ubuntu22.04] [--image-name db-gpt]" echo "USAGE: $0 [--base-image nvidia/cuda:12.1.0-runtime-ubuntu22.04] [--image-name db-gpt]"
echo " [-b|--base-image base image name] Base image name" echo " [-b|--base-image base image name] Base image name"
echo " [-n|--image-name image name] Current image name, default: db-gpt" echo " [-n|--image-name image name] Current image name, default: db-gpt"
echo " [--image-name-suffix image name suffix] Image name suffix"
echo " [-i|--pip-index-url pip index url] Pip index url, default: https://pypi.org/simple" echo " [-i|--pip-index-url pip index url] Pip index url, default: https://pypi.org/simple"
echo " [--language en or zh] You language, default: en" echo " [--language en or zh] You language, default: en"
echo " [--build-local-code true or false] Whether to use the local project code to package the image, default: true" echo " [--build-local-code true or false] Whether to use the local project code to package the image, default: true"
echo " [--load-examples true or false] Whether to load examples to default database default: true" echo " [--load-examples true or false] Whether to load examples to default database default: true"
echo " [--network network name] The network of docker build" echo " [--network network name] The network of docker build"
echo " [--install-mode mode name] Installation mode name, default: default, If you completely use openai's service, you can set the mode name to 'openai'" echo " [--install-mode mode name] Installation mode name, default: default, If you completely use openai's service, you can set the mode name to 'openai'"
echo " [-f|--dockerfile dockerfile] Dockerfile name, default: Dockerfile"
echo " [-h|--help] Usage message" echo " [-h|--help] Usage message"
} }
@ -46,6 +51,11 @@ while [[ $# -gt 0 ]]; do
shift # past argument shift # past argument
shift # past value shift # past value
;; ;;
--image-name-suffix)
IMAGE_NAME_SUFFIX="$2"
shift # past argument
shift # past value
;;
-i|--pip-index-url) -i|--pip-index-url)
PIP_INDEX_URL="$2" PIP_INDEX_URL="$2"
shift shift
@ -80,6 +90,11 @@ while [[ $# -gt 0 ]]; do
shift # past argument shift # past argument
shift # past value shift # past value
;; ;;
-f|--dockerfile)
DOCKERFILE="$2"
shift # past argument
shift # past value
;;
-h|--help) -h|--help)
help="true" help="true"
shift shift
@ -111,6 +126,10 @@ else
BASE_IMAGE=$IMAGE_NAME_ARGS BASE_IMAGE=$IMAGE_NAME_ARGS
fi fi
if [ -n "$IMAGE_NAME_SUFFIX" ]; then
IMAGE_NAME="$IMAGE_NAME-$IMAGE_NAME_SUFFIX"
fi
echo "Begin build docker image, base image: ${BASE_IMAGE}, target image name: ${IMAGE_NAME}" echo "Begin build docker image, base image: ${BASE_IMAGE}, target image name: ${IMAGE_NAME}"
docker build $BUILD_NETWORK \ docker build $BUILD_NETWORK \
@ -120,5 +139,5 @@ docker build $BUILD_NETWORK \
--build-arg BUILD_LOCAL_CODE=$BUILD_LOCAL_CODE \ --build-arg BUILD_LOCAL_CODE=$BUILD_LOCAL_CODE \
--build-arg LOAD_EXAMPLES=$LOAD_EXAMPLES \ --build-arg LOAD_EXAMPLES=$LOAD_EXAMPLES \
--build-arg DB_GPT_INSTALL_MODEL=$DB_GPT_INSTALL_MODEL \ --build-arg DB_GPT_INSTALL_MODEL=$DB_GPT_INSTALL_MODEL \
-f Dockerfile \ -f $DOCKERFILE \
-t $IMAGE_NAME $WORK_DIR/../../ -t $IMAGE_NAME $WORK_DIR/../../