How to deploy MPT with llama.cpp? #939
Unanswered
streetycat
asked this question in
Q&A
Replies: 3 comments 3 replies
-
How are you sending the POST? Can you share your |
Beta Was this translation helpful? Give feedback.
0 replies
-
Thank you, I have found a parameter( and update the command:
the responce: By the way, maybe the |
Beta Was this translation helpful? Give feedback.
3 replies
-
Definitely looks like prompt format issue |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I found that llama.cpp already supports mpt, I downloaded gguf from here, and it did load it with llama.cpp, but its return result looks bad.
I start the server as follow:
And I post the request as follow:
But, it works well on the demo page:

I worked well with llama in the same method, I don't know what's wrong happened.
Beta Was this translation helpful? Give feedback.
All reactions