model size mismatch
Xt117 opened this issue · 10 comments
Hi @Xt117 ,
would you mind provide more details regarding your settings? e.g., which checkpoints you are use? (clip / slowfast + clip)
which script you use?
thank.
okay, I am investigating this issue now.
@Xt117 , have found the reason, there is because I upload the mismatch checkpoint for pt+ft, sorry for the mistake, i have re-upload accordingly.
In the same google drive, with download link: https://drive.google.com/file/d/1gy8wKqA9gcYbk3tHewXX5qZ9SQAFhk6J/view?usp=drive_link.
The checkpoint size is ~150mb, can you please take a try and let me know?
Hi @Xt117 , i have uploaded the repo with a tmp
dir. can you manually create this tmp
under univtg
and try it again?
Thanks. Now it ran successful.
I used my own video to generate the top5 interval, but the results show 5 similar shots
how can i get the top5 different shots of total video like your visualization?
Otherwise, when i used the slowfast+clip finetune model, it got error like below, seems that the shape is mismatch
@Xt117 glad you can run successfully.
currently, the top similar shots are not under NMS, if you want to show five different shots, you should input these predicted windows into the nms function with a threshold. you can find my nms function in the utils i.e., https://github.com/showlab/UniVTG/blob/main/utils/temporal_nms.py
or implement it by yourself. I may update this detail later.
my current codes only support clip video features.
while the slowfast+clip finetune model is a more strong model, which inputs slowfast + clip features for video, thus we need to extract slowfast features based on https://github.com/linjieli222/HERO_Video_Feature_Extractor,
which are not included so far, I plan to include this in the next phase.
OK, I will try the nms function. Thanks.
Looking forward to your update.
close since have resolved the issues of mismatch model size.