-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]: different eval results compared to the results in paper #99
Comments
Hi @unicorneeee, thanks for your question. Could you share your scripts with us? Let me know if it helps! |
I use the run_infinitebench.sh to run the infinitebench,the code is as follow and I directly use the run_infinitebench.py without any change |
Hi @unicorneeee, we use |
Thanks for your answer! I will try it again! |
Hello! |
Hi @unicorneeee, I haven’t encountered this issue before. Llama3+ is already bf16 by default. https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-1048k/blob/main/config.json#L24 By the way, are you using flash_attn or our custom triton ops? |
Thanks for your answer! The model evaled in the paper is llama3-8B-262K, why this link is 1048K model? And I use the link in paper to download the 262K model, which I found that model this fp32, however the link is unavaible now? |
Hi @unicorneeee, apologies for the confusion—I just shared a random model as an example. You can refer to this configuration instead: https://huggingface.co/gradientai/Llama-3-8B-Instruct-262k/blob/main/config.json#L24. |
Hello! I have use the model downloaded by the link, and find the results is different with the results in paper:(especially in code debug and EN dia task) export TOKENIZERS_PARALLELISM=false |
Describe the issue
hello ,when I use the env as follows:
transformers:4.47.0
torch 2.3.0
triton 2.1.0
flash_attn 2.5.8
minference use the support_tp branch
however, I set the attn_type="hf" to evalute the infinitebench, there are different results compared to the results in paper:
results self-evalute (use llama3-8B-262K)
Llama-3-8B-Instruct-262k_hf,code_debug,24.62
Llama-3-8B-Instruct-262k_hf,math_find,18.00
Llama-3-8B-Instruct-262k_hf,longdialogue_qa_eng,0.50
Could U please provide the requirments of env when U test the infinitebench? thank you!
The text was updated successfully, but these errors were encountered: