mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-07 05:52:15 +00:00
openai[patch]: code cleaning (#17355)
h/t @tdene for finding cleanup op in #17047
This commit is contained in:
@@ -198,9 +198,11 @@ class BaseOpenAI(BaseLLM):
|
||||
)
|
||||
|
||||
client_params = {
|
||||
"api_key": values["openai_api_key"].get_secret_value()
|
||||
if values["openai_api_key"]
|
||||
else None,
|
||||
"api_key": (
|
||||
values["openai_api_key"].get_secret_value()
|
||||
if values["openai_api_key"]
|
||||
else None
|
||||
),
|
||||
"organization": values["openai_organization"],
|
||||
"base_url": values["openai_api_base"],
|
||||
"timeout": values["request_timeout"],
|
||||
@@ -257,9 +259,11 @@ class BaseOpenAI(BaseLLM):
|
||||
chunk.text,
|
||||
chunk=chunk,
|
||||
verbose=self.verbose,
|
||||
logprobs=chunk.generation_info["logprobs"]
|
||||
if chunk.generation_info
|
||||
else None,
|
||||
logprobs=(
|
||||
chunk.generation_info["logprobs"]
|
||||
if chunk.generation_info
|
||||
else None
|
||||
),
|
||||
)
|
||||
|
||||
async def _astream(
|
||||
@@ -283,9 +287,11 @@ class BaseOpenAI(BaseLLM):
|
||||
chunk.text,
|
||||
chunk=chunk,
|
||||
verbose=self.verbose,
|
||||
logprobs=chunk.generation_info["logprobs"]
|
||||
if chunk.generation_info
|
||||
else None,
|
||||
logprobs=(
|
||||
chunk.generation_info["logprobs"]
|
||||
if chunk.generation_info
|
||||
else None
|
||||
),
|
||||
)
|
||||
|
||||
def _generate(
|
||||
@@ -334,12 +340,16 @@ class BaseOpenAI(BaseLLM):
|
||||
choices.append(
|
||||
{
|
||||
"text": generation.text,
|
||||
"finish_reason": generation.generation_info.get("finish_reason")
|
||||
if generation.generation_info
|
||||
else None,
|
||||
"logprobs": generation.generation_info.get("logprobs")
|
||||
if generation.generation_info
|
||||
else None,
|
||||
"finish_reason": (
|
||||
generation.generation_info.get("finish_reason")
|
||||
if generation.generation_info
|
||||
else None
|
||||
),
|
||||
"logprobs": (
|
||||
generation.generation_info.get("logprobs")
|
||||
if generation.generation_info
|
||||
else None
|
||||
),
|
||||
}
|
||||
)
|
||||
else:
|
||||
@@ -395,12 +405,16 @@ class BaseOpenAI(BaseLLM):
|
||||
choices.append(
|
||||
{
|
||||
"text": generation.text,
|
||||
"finish_reason": generation.generation_info.get("finish_reason")
|
||||
if generation.generation_info
|
||||
else None,
|
||||
"logprobs": generation.generation_info.get("logprobs")
|
||||
if generation.generation_info
|
||||
else None,
|
||||
"finish_reason": (
|
||||
generation.generation_info.get("finish_reason")
|
||||
if generation.generation_info
|
||||
else None
|
||||
),
|
||||
"logprobs": (
|
||||
generation.generation_info.get("logprobs")
|
||||
if generation.generation_info
|
||||
else None
|
||||
),
|
||||
}
|
||||
)
|
||||
else:
|
||||
|
Reference in New Issue
Block a user