How to deploy MPT with llama.cpp? #939
Unanswered
streetycat
asked this question in
Q&A
Replies: 3 comments 3 replies
-
How are you sending the POST? Can you share your |
Beta Was this translation helpful? Give feedback.
0 replies
-
Thank you, I have found a parameter( and update the command:
the responce: By the way, maybe the |
Beta Was this translation helpful? Give feedback.
3 replies
-
Definitely looks like prompt format issue |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I found that llama.cpp already supports mpt, I downloaded gguf from here, and it did load it with llama.cpp, but its return result looks bad.
I start the server as follow:
And I post the request as follow:
But, it works well on the demo page:

I worked well with llama in the same method, I don't know what's wrong happened.
Beta Was this translation helpful? Give feedback.
All reactions