Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

is it possible to use llama,cpp with other neural networks? #158

Closed
dbpaul opened this issue Mar 15, 2023 · 2 comments
Closed

is it possible to use llama,cpp with other neural networks? #158

dbpaul opened this issue Mar 15, 2023 · 2 comments
Labels
enhancement New feature or request model Model specific

Comments

@dbpaul
Copy link

dbpaul commented Mar 15, 2023

I have no clue about this, but I saw that chatglm-6b was published, which should run on CPU with 16GB ram, albeit very slow.
https://huggingface.co/THUDM/chatglm-6b/tree/main

Would it be possible to substitute the llama model?

@v3ss0n
Copy link

v3ss0n commented Mar 15, 2023

if you are going to write the code in c base on ggml , yes.
Also please move to
https://github.com/ggerganov/llama.cpp/discussions

@gjmulder gjmulder added enhancement New feature or request model Model specific labels Mar 15, 2023
@Ayushk4
Copy link

Ayushk4 commented Mar 25, 2023

So far 10 different models are supported across 5 different architectures (including OpenAssistant and Open-Chat-Kit models) are supported by nolanoorg/cformers.

You can now interface with the models with just 3 lines of code from python.

from interface import AutoInference as AI
ai = AI('OpenAssistant/oasst-sft-1-pythia-12b')
x = ai.generate("<|prompter|>What's the Earth total population<|endoftext|><|assistant|>", num_tokens_to_generate=100); print(x['token_str'])

Generation speed is same as this repo (75 ms/token for 12B model on Macbook Pro)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request model Model specific
Projects
None yet
Development

No branches or pull requests

5 participants