langchain/libs
Yuwen Hu ba0dca46d7
community[minor]: Add IPEX-LLM BGE embedding support on both Intel CPU and GPU (#22226)
**Description:** [IPEX-LLM](https://github.com/intel-analytics/ipex-llm)
is a PyTorch library for running LLM on Intel CPU and GPU (e.g., local
PC with iGPU, discrete GPU such as Arc, Flex and Max) with very low
latency. This PR adds ipex-llm integrations to langchain for BGE
embedding support on both Intel CPU and GPU.
**Dependencies:** `ipex-llm`, `sentence-transformers`
**Contribution maintainer**: @Oscilloscope98 
**tests and docs**: 
- langchain/docs/docs/integrations/text_embedding/ipex_llm.ipynb
- langchain/docs/docs/integrations/text_embedding/ipex_llm_gpu.ipynb
-
langchain/libs/community/tests/integration_tests/embeddings/test_ipex_llm.py

---------

Co-authored-by: Shengsheng Huang <shannie.huang@gmail.com>
2024-06-03 12:37:10 -07:00
..
cli docs, cli[patch]: chat model template nit (#22294) 2024-05-29 20:53:58 +00:00
community community[minor]: Add IPEX-LLM BGE embedding support on both Intel CPU and GPU (#22226) 2024-06-03 12:37:10 -07:00
core core[patch]: RFC: Allow concatenation of messages with multi part content (#22002) 2024-06-03 09:46:40 -07:00
experimental [experimental][llms][ollama_functions] Update OllamaFunctions to send tool_calls attribute (#21625) 2024-05-29 09:38:33 -04:00
langchain Update parent_document_retriever.py about **kwargs (#22219) 2024-05-28 11:35:38 -07:00
partners infra: bump anthropic mypy 1 (#22373) 2024-06-03 08:21:55 -07:00
standard-tests core, partners: add token usage attribute to AIMessage (#21944) 2024-05-23 14:21:58 -04:00
text-splitters fix error message (#22437) 2024-06-03 15:48:26 +00:00