-
Notifications
You must be signed in to change notification settings - Fork 1.7k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Question] Does it support multi-gpu (intel ARC A770)?
question
Question about the usage
#3175
opened Mar 14, 2025 by
savvadesogle
A significant difference in answer quality between the model provided by the official and the model obtained by converting weights based on official documents
question
Question about the usage
#3174
opened Mar 14, 2025 by
FFchopon
[Bug] The Medusa model has some differences from the current official implementation
bug
Confirmed bugs
#3173
opened Mar 14, 2025 by
Songyanfei
[Question] How to get last layer hidden state of transformer model while transfer the model to MLC format?
question
Question about the usage
#3170
opened Mar 11, 2025 by
Jianshu1only
[Bug] App closes itself when initializing a model, DeepSeek-R1-Distill-Qwen-1.5B-q4f16_1-MLC
bug
Confirmed bugs
#3169
opened Mar 11, 2025 by
GTMssj
[Question] VLM inference on android
question
Question about the usage
#3167
opened Mar 10, 2025 by
amirvenus
[Bug] <iframe src="https://reach-vb-smollm2-structured-generation.static.hf.space" frameborder="0" width="850" height="450" ></iframe>
bug
Confirmed bugs
#3163
opened Mar 10, 2025 by
Jose1370
[Bug] /mlc-llm/3rdparty/tvm/src/runtime/relax_vm/attn_utils.h:712:19: error: no member named 'cl' in namespace 'tvm::runtime'
bug
Confirmed bugs
#3157
opened Mar 8, 2025 by
PineJuneYang
[Bug] Unable to convert weight: "PagedKVCache.attention_with_fused_qkv() missing 1 required positional argument: 'sm_scale'"
bug
Confirmed bugs
#3149
opened Mar 2, 2025 by
Kisaragi-ng
[Question] mlc-llm server cannot return correct logprobs
question
Question about the usage
#3142
opened Feb 19, 2025 by
kunxiongzhu
[Question] how to use function call
question
Question about the usage
#3141
opened Feb 19, 2025 by
tebie6
[Model Request] GLINER for entity recognisition
new-models
#3139
opened Feb 17, 2025 by
manasaniprashanth
[Bug] Gemma 2 models fail due to errors in tokenizer
bug
Confirmed bugs
#3138
opened Feb 17, 2025 by
julioasotodv
[Question] I followed the instructions to build for Orange Pi, but it seems outdated (ChatModule)
question
Question about the usage
#3134
opened Feb 16, 2025 by
LivingLinux
[Question] While waiting for the model's response on an Android phone, performing other operations may cause the phone to become unresponsive or reboot.
question
Question about the usage
#3131
opened Feb 13, 2025 by
yangshgetui
[Bug] mlc-llm server cannot return correct logprobs
bug
Confirmed bugs
#3130
opened Feb 13, 2025 by
kunxiongzhu
[Bug] Mistral-Nemo-Instruct-2407 The results were confused
bug
Confirmed bugs
#3120
opened Feb 7, 2025 by
fierceX
Very slow time to first token on ROCM
question
Question about the usage
#3119
opened Feb 5, 2025 by
Jyers
[Bug] Android app does not take input; 'user 'role' is not defined' error
bug
Confirmed bugs
#3117
opened Feb 4, 2025 by
afsara-ben
Previous Next
ProTip!
no:milestone will show everything without a milestone.