Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update max_new_tokens slider in web gui to support at least 16k #3835

Closed
marmelade500 opened this issue Sep 7, 2023 · 4 comments
Closed
Labels
enhancement New feature or request stale

Comments

@marmelade500
Copy link

maxnewtoken

@marmelade500 marmelade500 added the enhancement New feature or request label Sep 7, 2023
@Flanua
Copy link

Flanua commented Sep 7, 2023

I don't think there is AI models that support that yet. You can use RoPE scaling for that if your model supports it.

P.S: More info here:
ggml-org/llama.cpp#2054

@berkut1
Copy link
Contributor

berkut1 commented Sep 12, 2023

@Flanua I think the OP is talking about token generation (like long respond), not context.

If this is equal to llama.embedding_length, then some models support 5120 or even 16k.

@marmelade500
Copy link
Author

marmelade500 commented Sep 19, 2023

For anyone who wants a workaround, you can create a copy of the settings yaml, rename it to settings.yaml, and put your own custom settings in it. It will use those settings when the program is started.

@github-actions github-actions bot added the stale label Nov 1, 2023
Copy link

github-actions bot commented Nov 1, 2023

This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment.

@github-actions github-actions bot closed this as completed Nov 1, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request stale
Projects
None yet
Development

No branches or pull requests

3 participants