-
Notifications
You must be signed in to change notification settings - Fork 281
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] ValueError: Tokenizer class Qwen2Tokenizer does not exist or is not currently imported.
#1903
opened Jul 3, 2024 by
zhyncs
2 tasks done
[Bug] Using the turbomind engine, prompting more than 10k tokens will result in garbage output.
#1896
opened Jul 2, 2024 by
dafu-wu
2 tasks done
[Bug] CUDA runtime error: an illegal memory access was encountered when 8bit kv quant was enabled
#1895
opened Jul 1, 2024 by
aabbccddwasd
2 tasks done
[Bug] AttributeError: 'LlavaNextConfig' object has no attribute 'hidden_size'
#1868
opened Jun 27, 2024 by
zhaozeno
1 of 2 tasks
使用pipeline加载Qwen1.5-32B-Chat,tp=4,使用openai prompt格式提示其清洗中文但生成回复都是英文
#1864
opened Jun 26, 2024 by
Yang-bug-star
使用OpenAI format的输入得到的response要如何提取出回复文本,返回的response好像是分段的
#1863
opened Jun 26, 2024 by
Yang-bug-star
[Bug] Segmentation fault: address not mapped to object at address 0x2058
#1849
opened Jun 25, 2024 by
austingg
2 tasks done
[Bug] InternLM2MLP.forward() missing 1 required positional argument: 'im_mask'
#1847
opened Jun 25, 2024 by
jiangjingz
2 tasks done
[Feature] How to support bf16 when inferencing Internvl-chat
#1839
opened Jun 24, 2024 by
Leo-yang-1020
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.