How to compare the inference time?
Closed this issue · 4 comments
Gumpest commented
Hi, authors. I wonder how to present the efficiency via inference time.
Gumpest commented
Besides, I want to learn how to compute the CUDA time. Thanks a lot.
Yxxxb commented
Yxxxb commented
The overall batch size is 128, which is the same as LLaVA SFT stage. You could check the training Hyperparameters in the "Additional Implement Details" section of our paper's appendix.