Skip to content

Commit 950f9d0

Browse files
cli: fixed dead links to tools/main for cli and completion, fixed code owners
1 parent 4d5ae24 commit 950f9d0

File tree

7 files changed

+13
-9
lines changed

7 files changed

+13
-9
lines changed

CODEOWNERS

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -87,7 +87,8 @@
8787
/tests/ @ggerganov
8888
/tests/test-chat-.* @pwilkin
8989
/tools/batched-bench/ @ggerganov
90-
/tools/main/ @ggerganov
90+
/tools/cli/ @ngxson
91+
/tools/completion/ @ggerganov
9192
/tools/mtmd/ @ngxson
9293
/tools/perplexity/ @ggerganov
9394
/tools/quantize/ @ggerganov

README.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -313,7 +313,7 @@ The Hugging Face platform provides a variety of online tools for converting, qua
313313

314314
To learn more about model quantization, [read this documentation](tools/quantize/README.md)
315315

316-
## [`llama-cli`](tools/main)
316+
## [`llama-cli`](tools/cli)
317317

318318
#### A CLI tool for accessing and experimenting with most of `llama.cpp`'s functionality.
319319

@@ -525,7 +525,8 @@ To learn more about model quantization, [read this documentation](tools/quantize
525525
526526
## Other documentation
527527
528-
- [main (cli)](tools/main/README.md)
528+
- [cli](tools/cli/README.md)
529+
- [completion](tools/completion/README.md)
529530
- [server](tools/server/README.md)
530531
- [GBNF grammars](grammars/README.md)
531532

docs/development/HOWTO-add-model.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,8 @@ Adding a model requires few steps:
99
After following these steps, you can open PR.
1010

1111
Also, it is important to check that the examples and main ggml backends (CUDA, METAL, CPU) are working with the new architecture, especially:
12-
- [main](/tools/main/)
12+
- [cli](/tools/cli/)
13+
- [completion](/tools/completion/)
1314
- [imatrix](/tools/imatrix/)
1415
- [quantize](/tools/quantize/)
1516
- [server](/tools/server/)

grammars/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# GBNF Guide
22

3-
GBNF (GGML BNF) is a format for defining [formal grammars](https://en.wikipedia.org/wiki/Formal_grammar) to constrain model outputs in `llama.cpp`. For example, you can use it to force the model to generate valid JSON, or speak only in emojis. GBNF grammars are supported in various ways in `tools/main` and `tools/server`.
3+
GBNF (GGML BNF) is a format for defining [formal grammars](https://en.wikipedia.org/wiki/Formal_grammar) to constrain model outputs in `llama.cpp`. For example, you can use it to force the model to generate valid JSON, or speak only in emojis. GBNF grammars are supported in various ways in `tools/cli`, `tools/completion` and `tools/server`.
44

55
## Background
66

@@ -135,7 +135,7 @@ While semantically correct, the syntax `x? x? x?.... x?` (with N repetitions) ma
135135
You can use GBNF grammars:
136136

137137
- In [llama-server](../tools/server)'s completion endpoints, passed as the `grammar` body field
138-
- In [llama-cli](../tools/main), passed as the `--grammar` & `--grammar-file` flags
138+
- In [llama-cli](../tools/cli) and [llama-completion](../tools/completion), passed as the `--grammar` & `--grammar-file` flags
139139
- With [test-gbnf-validator](../tests/test-gbnf-validator.cpp), to test them against strings.
140140

141141
## JSON Schemas → GBNF
@@ -145,7 +145,7 @@ You can use GBNF grammars:
145145
- In [llama-server](../tools/server):
146146
- For any completion endpoints, passed as the `json_schema` body field
147147
- For the `/chat/completions` endpoint, passed inside the `response_format` body field (e.g. `{"type", "json_object", "schema": {"items": {}}}` or `{ type: "json_schema", json_schema: {"schema": ...} }`)
148-
- In [llama-cli](../tools/main), passed as the `--json` / `-j` flag
148+
- In [llama-cli](../tools/cli) and [llama-completion](../tools/completion), passed as the `--json` / `-j` flag
149149
- To convert to a grammar ahead of time:
150150
- in CLI, with [examples/json_schema_to_grammar.py](../examples/json_schema_to_grammar.py)
151151
- in JavaScript with [json-schema-to-grammar.mjs](../tools/server/public_legacy/json-schema-to-grammar.mjs) (this is used by the [server](../tools/server)'s Web UI)

tools/cli/README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
TODO

tools/completion/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# llama.cpp/tools/main
1+
# llama.cpp/tools/completion
22

33
This example program allows you to use various LLaMA language models easily and efficiently. It is specifically designed to work with the [llama.cpp](https://github.com/ggml-org/llama.cpp) project, which provides a plain C/C++ implementation with optional 4-bit quantization support for faster, lower memory inference, and is optimized for desktop CPUs. This program can be used to perform various inference tasks with LLaMA models, including generating text based on user-provided prompts and chat-like interactions with reverse prompts.
44

tools/llama-bench/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ Each test is repeated the number of times given by `-r`, and the results are ave
8080

8181
Using the `-d <n>` option, each test can be run at a specified context depth, prefilling the KV cache with `<n>` tokens.
8282

83-
For a description of the other options, see the [main example](../main/README.md).
83+
For a description of the other options, see the [completion example](../completion/README.md).
8484

8585
> [!NOTE]
8686
> The measurements with `llama-bench` do not include the times for tokenization and for sampling.

0 commit comments

Comments
 (0)