Update README.md

This commit is contained in:
Andriy Mulyar 2023-03-28 16:31:53 -04:00 committed by GitHub
parent bc7b2d256e
commit 83f80b5a78

View File

@ -11,7 +11,7 @@
# Try it yourself # Try it yourself
You can download pre-compiled LLaMa C++ Interactive Chat binaries here: You can download pre-compiled GPT4ALL Interactive Chat binaries here:
- [OSX](https://s3.amazonaws.com/static.nomic.ai/gpt4all/models/gpt4all-lora-quantized-OSX-m1) - [OSX](https://s3.amazonaws.com/static.nomic.ai/gpt4all/models/gpt4all-lora-quantized-OSX-m1)
- [Intel/Windows]() - [Intel/Windows]()
@ -20,7 +20,7 @@ and the model
Place the binary and quantized model in the same directory and start chatting! Place the binary and quantized model in the same directory and start chatting!
To compile for custom hardware, see our fork of the [Alpaca C++](https://github.com/zanussbaum/alpaca.cpp) repo. To compile for custom hardware, see our fork of the [Alpaca C++](https://github.com/zanussbaum/gpt4all.cpp) repo.
# Reproducibility # Reproducibility