Skip to content

Issues: InternLM/lmdeploy

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

是否支持VLM 输出logprobs
#3243 opened Mar 12, 2025 by hxyghostor
使用batch推理进行多轮对话
#3238 opened Mar 11, 2025 by 2kxx
[Bug] NPU下Qwen2.5-VL推理报错
#3237 opened Mar 10, 2025 by tcye
3 tasks done
[Bug] no chat-template for QwQ 32B AWQ
#3224 opened Mar 6, 2025 by HuChundong
3 tasks
LMDeploy Lite vs LLM Compressor
#3199 opened Mar 1, 2025 by radna0
[Feature] accelerate QwenEncoder
#3185 opened Feb 25, 2025 by abc8350712
多轮对话的batch推理
#3184 opened Feb 25, 2025 by wzczc
ProTip! What’s not been updated in a month: updated:<2025-02-12.