File tree 1 file changed +3
-2
lines changed
1 file changed +3
-2
lines changed Original file line number Diff line number Diff line change 1
1
# llama.cpp/example/server
2
2
3
- This example demonstrates a simple HTTP API server to interact with llama.cpp.
3
+ This example demonstrates a simple HTTP API server and a simple web front end to interact with llama.cpp.
4
4
5
5
Command line options:
6
6
@@ -21,6 +21,7 @@ Command line options:
21
21
- ` -to N ` , ` --timeout N ` : Server read/write timeout in seconds. Default ` 600 ` .
22
22
- ` --host ` : Set the hostname or ip address to listen. Default ` 127.0.0.1 ` .
23
23
- ` --port ` : Set the port to listen. Default: ` 8080 ` .
24
+ - ` --public ` : path from which to serve static files (default examples/server/public)
24
25
- ` --embedding ` : Enable embedding extraction, Default: disabled.
25
26
26
27
## Build
@@ -59,7 +60,7 @@ server.exe -m models\7B\ggml-model.bin -c 2048
59
60
```
60
61
61
62
The above command will start a server that by default listens on ` 127.0.0.1:8080 ` .
62
- You can consume the endpoints with Postman or NodeJS with axios library.
63
+ You can consume the endpoints with Postman or NodeJS with axios library. You can visit the web front end at the same url.
63
64
64
65
## Testing with CURL
65
66
You can’t perform that action at this time.
0 commit comments