Skip to content

Commit 334637e

Browse files
authored
common : change default parameters to pre-#1126 (#1223)
1 parent dd7eff5 commit 334637e

File tree

2 files changed

+25
-25
lines changed

2 files changed

+25
-25
lines changed

Diff for: examples/common.h

+13-13
Original file line numberDiff line numberDiff line change
@@ -17,26 +17,26 @@
1717
struct gpt_params {
1818
int32_t seed = -1; // RNG seed
1919
int32_t n_threads = std::min(4, (int32_t) std::thread::hardware_concurrency());
20-
int32_t n_predict = 128; // new tokens to predict
20+
int32_t n_predict = -1; // new tokens to predict
2121
int32_t n_parts = -1; // amount of model parts (-1 = determine from model dimensions)
2222
int32_t n_ctx = 512; // context size
2323
int32_t n_batch = 512; // batch size for prompt processing (must be >=32 to use BLAS)
2424
int32_t n_keep = 0; // number of tokens to keep from initial prompt
2525

2626
// sampling parameters
2727
std::unordered_map<llama_token, float> logit_bias; // logit bias for specific tokens
28-
int32_t top_k = 0; // <= 0 to use vocab size
29-
float top_p = 1.0f; // 1.0 = disabled
30-
float tfs_z = 1.0f; // 1.0 = disabled
31-
float typical_p = 1.0f; // 1.0 = disabled
32-
float temp = 1.0f; // 1.0 = disabled
33-
float repeat_penalty = 1.0f; // 1.0 = disabled
34-
int32_t repeat_last_n = -1; // last n tokens to penalize (0 = disable penalty, -1 = context size)
35-
float frequency_penalty = 0.0f; // 0.0 = disabled
36-
float presence_penalty = 0.0f; // 0.0 = disabled
37-
int mirostat = 0; // 0 = disabled, 1 = mirostat, 2 = mirostat 2.0
38-
float mirostat_tau = 5.0f; // target entropy
39-
float mirostat_eta = 0.1f; // learning rate
28+
int32_t top_k = 40; // <= 0 to use vocab size
29+
float top_p = 0.95f; // 1.0 = disabled
30+
float tfs_z = 1.00f; // 1.0 = disabled
31+
float typical_p = 1.00f; // 1.0 = disabled
32+
float temp = 0.80f; // 1.0 = disabled
33+
float repeat_penalty = 1.10f; // 1.0 = disabled
34+
int32_t repeat_last_n = 64; // last n tokens to penalize (0 = disable penalty, -1 = context size)
35+
float frequency_penalty = 0.00f; // 0.0 = disabled
36+
float presence_penalty = 0.00f; // 0.0 = disabled
37+
int mirostat = 0; // 0 = disabled, 1 = mirostat, 2 = mirostat 2.0
38+
float mirostat_tau = 5.00f; // target entropy
39+
float mirostat_eta = 0.10f; // learning rate
4040

4141
std::string model = "models/lamma-7B/ggml-model.bin"; // model path
4242
std::string prompt = "";

Diff for: examples/main/main.cpp

+12-12
Original file line numberDiff line numberDiff line change
@@ -387,19 +387,19 @@ int main(int argc, char ** argv) {
387387

388388
if ((int) embd_inp.size() <= n_consumed && !is_interacting) {
389389
// out of user input, sample next token
390-
const float temp = params.temp;
391-
const int32_t top_k = params.top_k <= 0 ? llama_n_vocab(ctx) : params.top_k;
392-
const float top_p = params.top_p;
393-
const float tfs_z = params.tfs_z;
394-
const float typical_p = params.typical_p;
395-
const int32_t repeat_last_n = params.repeat_last_n < 0 ? n_ctx : params.repeat_last_n;
396-
const float repeat_penalty = params.repeat_penalty;
397-
const float alpha_presence = params.presence_penalty;
390+
const float temp = params.temp;
391+
const int32_t top_k = params.top_k <= 0 ? llama_n_vocab(ctx) : params.top_k;
392+
const float top_p = params.top_p;
393+
const float tfs_z = params.tfs_z;
394+
const float typical_p = params.typical_p;
395+
const int32_t repeat_last_n = params.repeat_last_n < 0 ? n_ctx : params.repeat_last_n;
396+
const float repeat_penalty = params.repeat_penalty;
397+
const float alpha_presence = params.presence_penalty;
398398
const float alpha_frequency = params.frequency_penalty;
399-
const int mirostat = params.mirostat;
400-
const float mirostat_tau = params.mirostat_tau;
401-
const float mirostat_eta = params.mirostat_eta;
402-
const bool penalize_nl = params.penalize_nl;
399+
const int mirostat = params.mirostat;
400+
const float mirostat_tau = params.mirostat_tau;
401+
const float mirostat_eta = params.mirostat_eta;
402+
const bool penalize_nl = params.penalize_nl;
403403

404404
// optionally save the session on first sample (for faster prompt loading next time)
405405
if (!path_session.empty() && need_to_save_session) {

0 commit comments

Comments
 (0)