-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support internlm2 #1392
Support internlm2 #1392
Conversation
@PaulX1029 Hi, have you rebuilt and reinstalled tensorrt-llm? You can find the installation location by |
@RunningLeon 请问您是采用什么build的方式,我是从pip安装的trtllm,我想要跟您对齐build方式重新进行 |
|
@RunningLeon 很感谢你的工作!想问下internvl-1.5的internlm2-20b 网络跟普通internlm2-20b有什么区别吗,我用了PR里的转换脚本转出来之后都是乱码。 |
@RunningLeon hi,we use lora finetuned the internlm2 model. Now we can convert the base model to llama, but not lora part, we tried to change the code of InternLM/tools/convert2llame.py to transfer lora to llama style, but did not work. Is there any other tools could work for lora? |
@nv-guomingz Hi, sorry to bother you, but when will this PR be merged? Do I need to fix the conflicts? |
cpp/tensorrt_llm/kernels/decoderMaskedMultiheadAttentionUtils.h
Outdated
Show resolved
Hide resolved
@nv-guomingz hi, the conflicts with main branch are resolved. Looking forward to your review comments. THX. |
@RunningLeon 请问下为什么internlm2需要单独一个convert_checkpoint.py呢?而不是复用llama的convert_checkpoint.py,internlm是直接使用llama的convert_checkpoint.py |
hi, internlm2 W_qkv是在一起的,其次一些参数命名是和llama没有对齐的。因而没法直接使用llama的convert_checkpoint.py |
Thank you for this explanation! |
Hi @RunningLeon sorry for late response due to internal task priority. |
@nv-guomingz Done. Hope merging with main is OK. |
Thanks @RunningLeon. Could u please rebase your commits into one single commit? That would be more easy for further integration. |
5a8ee31
to
94f57cc
Compare
Hi @RunningLeon I've managed to file the merge request in our internal repo and testing is on-going. |
@RunningLeon Internlm2 had been added into today's update. |
This PR supports the conversion of internlm2 from hf to trt-llm checkpoints with :