File tree Expand file tree Collapse file tree 2 files changed +25
-0
lines changed Expand file tree Collapse file tree 2 files changed +25
-0
lines changed Original file line number Diff line number Diff line change @@ -4,6 +4,15 @@ All notable changes to this project will be documented in this file.
4
4
The format is based on [ Keep a Changelog] ( https://keepachangelog.com/en/1.0.0/ ) ,
5
5
and this project adheres to [ Semantic Versioning] ( https://semver.org/spec/v2.0.0.html ) .
6
6
7
+ ## [ 1.2.0] - 2023-07-06
8
+
9
+ ### Added
10
+ - Add server example to the build
11
+ - Add documentation on how to use the webinterface
12
+
13
+ ### Fixed
14
+ - Fix automatic update of the submodules
15
+
7
16
## [ 1.1.0] - 2023-07-03
8
17
9
18
### Added
Original file line number Diff line number Diff line change @@ -103,6 +103,22 @@ You can now chat with the model:
103
103
--interactive
104
104
```
105
105
106
+ ### Chat via Webinterface
107
+
108
+ You can start llama.cpp as a webserver:
109
+
110
+ ``` PowerShell
111
+ ./vendor/llama.cpp/build/bin/Release/server `
112
+ --model "./vendor/llama.cpp/models/open-llama-7B-open-instruct.ggmlv3.q4_K_M.bin" `
113
+ --ctx-size 2048 `
114
+ --threads 16 `
115
+ --n-gpu-layers 32
116
+ ```
117
+
118
+ And then access llama.cpp via the webinterface at:
119
+
120
+ * http://localhost:8080/
121
+
106
122
### Measure model perplexity
107
123
108
124
Execute the following to measure the perplexity of the GGML formatted model:
You can’t perform that action at this time.
0 commit comments