Replies: 1 comment
-
Did you make some progress figuring out the proper chat template for vllm and deepseek ? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have tested both the llama.cpp and vllm with deepseek and there seems to be couple of big differences that I do not properly understand or have just configured in a way that causes differences
In vllm case I am for example sure whether I use the correct jinja file as a template parameter and could that be the reason why output is mix of chinese and english text. (--chat-template ./template_chatglm.jinja) Or would there be some option that would allow the webui to launch immediately before loading the whole model to GPU's vram.
Beta Was this translation helpful? Give feedback.
All reactions