From 59da9399cd865ee29c3dec3614a081f6eb075e77 Mon Sep 17 00:00:00 2001 From: setzer22 Date: Tue, 14 Mar 2023 13:04:45 +0100 Subject: [PATCH] Update README.md --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index df2021fc..ffe86e35 100644 --- a/README.md +++ b/README.md @@ -45,7 +45,8 @@ cargo run --release -- -m /data/Llama/LLaMA/7B/ggml-model-q4_0.bin -p "Tell me h Contributions welcome! Here's a few pressing issues: - [ ] The code only sets the right CFLAGS on Linux. The `build.rs` script in - `ggml_raw` needs to be fixed. + `ggml_raw` needs to be fixed, so inference *will be very slow on every + other OS*. - [ ] The quantization code has not been ported (yet). You can still use the quantized models with llama.cpp. - [ ] The code needs to be "library"-fied. It is nice as a showcase binary, but