Skip to content

Commit 8567c76

Browse files
jessejohnsonJesse Johnson
and
Jesse Johnson
authored
Update server instructions for web front end (ggml-org#2103)
Co-authored-by: Jesse Johnson <[email protected]>
1 parent 924dd22 commit 8567c76

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

examples/server/README.md

+3-2
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# llama.cpp/example/server
22

3-
This example demonstrates a simple HTTP API server to interact with llama.cpp.
3+
This example demonstrates a simple HTTP API server and a simple web front end to interact with llama.cpp.
44

55
Command line options:
66

@@ -21,6 +21,7 @@ Command line options:
2121
- `-to N`, `--timeout N`: Server read/write timeout in seconds. Default `600`.
2222
- `--host`: Set the hostname or ip address to listen. Default `127.0.0.1`.
2323
- `--port`: Set the port to listen. Default: `8080`.
24+
- `--public`: path from which to serve static files (default examples/server/public)
2425
- `--embedding`: Enable embedding extraction, Default: disabled.
2526

2627
## Build
@@ -59,7 +60,7 @@ server.exe -m models\7B\ggml-model.bin -c 2048
5960
```
6061

6162
The above command will start a server that by default listens on `127.0.0.1:8080`.
62-
You can consume the endpoints with Postman or NodeJS with axios library.
63+
You can consume the endpoints with Postman or NodeJS with axios library. You can visit the web front end at the same url.
6364

6465
## Testing with CURL
6566

0 commit comments

Comments
 (0)