mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-08 22:42:05 +00:00
all: test 3.13 ci (#27197)
Co-authored-by: Bagatur <baskaryan@gmail.com> Co-authored-by: Bagatur <22008038+baskaryan@users.noreply.github.com>
This commit is contained in:
@@ -58,11 +58,11 @@ def _create_retry_decorator(embeddings: OpenAIEmbeddings) -> Callable[[Any], Any
|
||||
max=embeddings.retry_max_seconds,
|
||||
),
|
||||
retry=(
|
||||
retry_if_exception_type(openai.error.Timeout)
|
||||
| retry_if_exception_type(openai.error.APIError)
|
||||
| retry_if_exception_type(openai.error.APIConnectionError)
|
||||
| retry_if_exception_type(openai.error.RateLimitError)
|
||||
| retry_if_exception_type(openai.error.ServiceUnavailableError)
|
||||
retry_if_exception_type(openai.error.Timeout) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.APIError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.APIConnectionError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.RateLimitError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.ServiceUnavailableError) # type: ignore[attr-defined]
|
||||
),
|
||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
||||
)
|
||||
@@ -85,11 +85,11 @@ def _async_retry_decorator(embeddings: OpenAIEmbeddings) -> Any:
|
||||
max=embeddings.retry_max_seconds,
|
||||
),
|
||||
retry=(
|
||||
retry_if_exception_type(openai.error.Timeout)
|
||||
| retry_if_exception_type(openai.error.APIError)
|
||||
| retry_if_exception_type(openai.error.APIConnectionError)
|
||||
| retry_if_exception_type(openai.error.RateLimitError)
|
||||
| retry_if_exception_type(openai.error.ServiceUnavailableError)
|
||||
retry_if_exception_type(openai.error.Timeout) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.APIError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.APIConnectionError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.RateLimitError) # type: ignore[attr-defined]
|
||||
| retry_if_exception_type(openai.error.ServiceUnavailableError) # type: ignore[attr-defined]
|
||||
),
|
||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
||||
)
|
||||
@@ -110,7 +110,7 @@ def _check_response(response: dict, skip_empty: bool = False) -> dict:
|
||||
if any(len(d["embedding"]) == 1 for d in response["data"]) and not skip_empty:
|
||||
import openai
|
||||
|
||||
raise openai.error.APIError("OpenAI API returned an empty embedding")
|
||||
raise openai.error.APIError("OpenAI API returned an empty embedding") # type: ignore[attr-defined]
|
||||
return response
|
||||
|
||||
|
||||
@@ -357,7 +357,7 @@ class OpenAIEmbeddings(BaseModel, Embeddings):
|
||||
**client_params
|
||||
).embeddings
|
||||
elif not values.get("client"):
|
||||
values["client"] = openai.Embedding
|
||||
values["client"] = openai.Embedding # type: ignore[attr-defined]
|
||||
else:
|
||||
pass
|
||||
return values
|
||||
@@ -390,7 +390,7 @@ class OpenAIEmbeddings(BaseModel, Embeddings):
|
||||
"Please install it with `pip install openai`."
|
||||
)
|
||||
|
||||
openai.proxy = {
|
||||
openai.proxy = { # type: ignore[attr-defined]
|
||||
"http": self.openai_proxy,
|
||||
"https": self.openai_proxy,
|
||||
} # type: ignore[assignment]
|
||||
|
Reference in New Issue
Block a user