-
Notifications
You must be signed in to change notification settings - Fork 212
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LongVILA - Base LLM not declared in the Paper? #152
Comments
LongVILA was originally trained with LLama3 (so as the releaesd models) and recently updated to Qwen2 backbone. @yukang2017 can help confirm |
@Lyken17 thank you Ligeng for the prompt reply. @yukang2017 would love to have the Qwen2 version on huggingface, so we can reproduce the results in the submission. thanks! |
Thanks for your interests in our work. We are waiting for legal permission from NVIDIA to release qwen2 models. It should be approved this week. I will let you know here when it is release on hf. Thanks for your patience. |
@liyucheng09 Updated results is in the paper. Paper: https://arxiv.org/pdf/2408.10188 |
@yukang2017 Hi Yukang, thanks for updating the latest model. Will try it soon. Close as solved. |
@yukang2017 Hi Yukang, the repo seems to be largely updated since the NVILA release. the old command to reproduce the results in the paper is not valid anymore. Any ideas how to run the experiments in the LongVILA paper? |
@yukang2017 @zhijian-liu Hi, it seems this repo is largely broken after the NVILA relase. The Error see below.
|
Hi @liyucheng09 , Sorry for the bug. I fix it in this pr. You can try this repo, https://github.com/yukang2017/VILA/tree/main, before it was merged. Regards, |
@yukang2017 many thanks, will try it very soon. |
@yukang2017 Hi Yukang, can you help to confirm the
I cannot reproduce the results in the paper, and I suspect it's due to a wrong conv template. |
@Lyken17 Hi Yukang, one more small question on reproduction. Should I specify |
Hi @liyucheng09 , I think we do not need to set conv_template when use vila-run evaluation. It can find the conv template from the tokenizer config json. I did not set add_newline_token before. What are the results do yo get? Did you follow the instructions here for evaluation? |
I'd like to ask the base LLM of the following LongVILA checkpoint:
Efficient-Large-Model/Llama-3-LongVILA-8B-128Frames
Efficient-Large-Model/Llama-3-LongVILA-8B-256Frames
Efficient-Large-Model/Llama-3-LongVILA-8B-512Frames
This was named with
Llama-3
, however, as quote from the paper:It is unclear what's the exact base LLM of LongVILA.
The text was updated successfully, but these errors were encountered: