Skip to content

Commit

Permalink
fix: default to llm-cli for cargo run
Browse files Browse the repository at this point in the history
  • Loading branch information
philpax committed May 6, 2023
1 parent b1a5c68 commit 14cd303
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 2 deletions.
1 change: 1 addition & 0 deletions Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ members = [
"binaries/*"
]
resolver = "2"
default-members = ["binaries/llm-cli"]

[workspace.package]
repository = "https://github.com/rustformers/llm"
Expand Down
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ The resulting binary will be at `target/release/llm[.exe]`.
It can also be run directly through Cargo, using

```shell
cargo run --release --bin llm -- <ARGS>
cargo run --release -- <ARGS>
```

This is useful for development.
Expand Down Expand Up @@ -113,7 +113,7 @@ is compatible with ggml. To achieve this, follow the steps outlined below:
python3 scripts/convert-pth-to-ggml.py /path/to/your/models/7B/ 1

# Quantize the model to 4-bit ggml format
cargo run --bin llm llama quantize /path/to/your/models/7B/ggml-model-f16.bin /path/to/your/models/7B/ggml-model-q4_0.bin q4_0
cargo run --release llama quantize /path/to/your/models/7B/ggml-model-f16.bin /path/to/your/models/7B/ggml-model-q4_0.bin q4_0
```

> **Note**
Expand Down

0 comments on commit 14cd303

Please sign in to comment.