We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent c07e87f commit 01f37edCopy full SHA for 01f37ed
examples/run/README.md
@@ -3,11 +3,10 @@
3
The purpose of this example is to demonstrate a minimal usage of llama.cpp for running models.
4
5
```bash
6
-llama-run granite-code
+llama-run granite3-moe
7
```
8
9
10
-llama-run -h
11
Description:
12
Runs a llm
13
@@ -17,7 +16,7 @@ Usage:
17
16
Options:
18
-c, --context-size <value>
19
Context size (default: 2048)
20
- -n, --ngl <value>
+ -n, -ngl, --ngl <value>
21
Number of GPU layers (default: 0)
22
--temp <value>
23
Temperature (default: 0.8)
0 commit comments