mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-02 03:26:17 +00:00
community[major], core[patch], langchain[patch], experimental[patch]: Create langchain-community (#14463)
Moved the following modules to new package langchain-community in a backwards compatible fashion: ``` mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community ``` Moved the following to core ``` mv langchain/langchain/utils/json_schema.py core/langchain_core/utils mv langchain/langchain/utils/html.py core/langchain_core/utils mv langchain/langchain/utils/strings.py core/langchain_core/utils cat langchain/langchain/utils/env.py >> core/langchain_core/utils/env.py rm langchain/langchain/utils/env.py ``` See .scripts/community_split/script_integrations.sh for all changes
This commit is contained in:
@@ -0,0 +1,57 @@
|
||||
"""Test Xinference wrapper."""
|
||||
import time
|
||||
from typing import AsyncGenerator, Tuple
|
||||
|
||||
import pytest_asyncio
|
||||
|
||||
from langchain_community.llms import Xinference
|
||||
|
||||
|
||||
@pytest_asyncio.fixture
|
||||
async def setup() -> AsyncGenerator[Tuple[str, str], None]:
|
||||
import xoscar as xo
|
||||
from xinference.deploy.supervisor import start_supervisor_components
|
||||
from xinference.deploy.utils import create_worker_actor_pool
|
||||
from xinference.deploy.worker import start_worker_components
|
||||
|
||||
pool = await create_worker_actor_pool(
|
||||
f"test://127.0.0.1:{xo.utils.get_next_port()}"
|
||||
)
|
||||
print(f"Pool running on localhost:{pool.external_address}")
|
||||
|
||||
endpoint = await start_supervisor_components(
|
||||
pool.external_address, "127.0.0.1", xo.utils.get_next_port()
|
||||
)
|
||||
await start_worker_components(
|
||||
address=pool.external_address, supervisor_address=pool.external_address
|
||||
)
|
||||
|
||||
# wait for the api.
|
||||
time.sleep(3)
|
||||
async with pool:
|
||||
yield endpoint, pool.external_address
|
||||
|
||||
|
||||
def test_xinference_llm_(setup: Tuple[str, str]) -> None:
|
||||
from xinference.client import RESTfulClient
|
||||
|
||||
endpoint, _ = setup
|
||||
|
||||
client = RESTfulClient(endpoint)
|
||||
|
||||
model_uid = client.launch_model(
|
||||
model_name="vicuna-v1.3", model_size_in_billions=7, quantization="q4_0"
|
||||
)
|
||||
|
||||
llm = Xinference(server_url=endpoint, model_uid=model_uid)
|
||||
|
||||
answer = llm(prompt="Q: What food can we try in the capital of France? A:")
|
||||
|
||||
assert isinstance(answer, str)
|
||||
|
||||
answer = llm(
|
||||
prompt="Q: where can we visit in the capital of France? A:",
|
||||
generate_config={"max_tokens": 1024, "stream": True},
|
||||
)
|
||||
|
||||
assert isinstance(answer, str)
|
Reference in New Issue
Block a user