-
Notifications
You must be signed in to change notification settings - Fork 254
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] Space is incorrectly removed from start of generated text for
/v1/completion
endpoint
#1743
opened Jun 8, 2024 by
josephrocca
2 tasks done
[Docs] Guidance on setting
num_tokens_per_iter
and max_prefill_iters
to optimal values
#1740
opened Jun 8, 2024 by
josephrocca
[Bug]
detokenize_incrementally
: OverflowError: out of range integral type conversion attempted
#1739
opened Jun 7, 2024 by
josephrocca
2 tasks done
[Bug] CUDA OOM during calibration even with 5x 4090s? Falling back to
--device cpu
also fails (with different error)
#1729
opened Jun 7, 2024 by
josephrocca
2 tasks done
High GPU memory for running InternVL-Chat-V1-5-AWQ
awaiting response
#1728
opened Jun 7, 2024 by
tairen99
[Bug] Mini-InternVL1.5-4B does not suceessfully initialized.
#1721
opened Jun 6, 2024 by
cydiachen
1 of 2 tasks
[Bug] internlm2-chat-1_8b模型使用4bit KV量化的时候找不到key_stats.pth
awaiting response
#1720
opened Jun 6, 2024 by
jxfruit
2 tasks
[Bug] Why does prefix caching change the generated content
#1719
opened Jun 5, 2024 by
DayDayupupupup
1 of 2 tasks
[Feature] lmdeploy通过命令行可以启动一个gradio应用,这个gradio的应用是不是可以给用户提供UI修改的方法?
awaiting response
#1710
opened Jun 5, 2024 by
kaiwang0112006
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.