mirror of
https://github.com/hwchase17/langchain.git
synced 2025-10-12 20:42:25 +00:00
Reopened as a personal repo outside the organization. ## Description - Naver HyperCLOVA X community package - Add chat model & embeddings - Add unit test & integration test - Add chat model & embeddings docs - I changed partner package(https://github.com/langchain-ai/langchain/pull/24252) to community package on this PR - Could this embeddings(https://github.com/langchain-ai/langchain/pull/21890) be deprecated? We are trying to replace it with embedding model(**ClovaXEmbeddings**) in this PR. Twitter handle: None. (if needed, contact with joonha.jeon@navercorp.com) --- you can check our previous discussion below: > one question on namespaces - would it make sense to have these in .clova namespaces instead of .naver? I would like to keep it as is, unless it is essential to unify the package name. (ClovaX is a branding for the model, and I plan to add other models and components. They need to be managed as separate classes.) > also, could you clarify the difference between ClovaEmbeddings and ClovaXEmbeddings? There are 3 models that are being serviced by embedding, and all are supported in the current PR. In addition, all the functionality of CLOVA Studio that serves actual models, such as distinguishing between test apps and service apps, is supported. The existing PR does not support this content because it is hard-coded. --------- Co-authored-by: Erick Friis <erick@langchain.dev> Co-authored-by: Vadym Barda <vadym@langchain.dev>
72 lines
2.0 KiB
Python
72 lines
2.0 KiB
Python
"""Test ChatNaver chat model."""
|
|
|
|
from langchain_core.messages import AIMessage, AIMessageChunk
|
|
|
|
from langchain_community.chat_models import ChatClovaX
|
|
|
|
|
|
def test_stream() -> None:
|
|
"""Test streaming tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
for token in llm.stream("I'm Clova"):
|
|
assert isinstance(token, AIMessageChunk)
|
|
assert isinstance(token.content, str)
|
|
|
|
|
|
async def test_astream() -> None:
|
|
"""Test streaming tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
async for token in llm.astream("I'm Clova"):
|
|
assert isinstance(token, AIMessageChunk)
|
|
assert isinstance(token.content, str)
|
|
|
|
|
|
async def test_abatch() -> None:
|
|
"""Test streaming tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
result = await llm.abatch(["I'm Clova", "I'm not Clova"])
|
|
for token in result:
|
|
assert isinstance(token, AIMessage)
|
|
assert isinstance(token.content, str)
|
|
|
|
|
|
async def test_abatch_tags() -> None:
|
|
"""Test batch tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
result = await llm.abatch(["I'm Clova", "I'm not Clova"], config={"tags": ["foo"]})
|
|
for token in result:
|
|
assert isinstance(token, AIMessage)
|
|
assert isinstance(token.content, str)
|
|
|
|
|
|
def test_batch() -> None:
|
|
"""Test batch tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
result = llm.batch(["I'm Clova", "I'm not Clova"])
|
|
for token in result:
|
|
assert isinstance(token, AIMessage)
|
|
assert isinstance(token.content, str)
|
|
|
|
|
|
async def test_ainvoke() -> None:
|
|
"""Test invoke tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
result = await llm.ainvoke("I'm Clova", config={"tags": ["foo"]})
|
|
assert isinstance(result, AIMessage)
|
|
assert isinstance(result.content, str)
|
|
|
|
|
|
def test_invoke() -> None:
|
|
"""Test invoke tokens from ChatClovaX."""
|
|
llm = ChatClovaX()
|
|
|
|
result = llm.invoke("I'm Clova", config=dict(tags=["foo"]))
|
|
assert isinstance(result, AIMessage)
|
|
assert isinstance(result.content, str)
|