zhiyuanyou/SAFECount

How different is this paper

Closed this issue · 3 comments

https://arxiv.org/pdf/2211.08217v1.pdf

@zhiyuanyou could you read through this once. author has explained the logic .
for attention feature they are using I think Transformer encoder layer . Later part not clear to me.
So wanted to understand if there is big or minor difference between your paper and this new one

Well, I had read this paper in the past. I think this paper proposes a better method considering its outstanding performance.

@zhiyuanyou any small help I can get in implementing its Object protoptype extraction part, It appears analogous to your Safe count block logic
After reading through this paper , it appears they use Resnet 50 as backbone followed by Multiheaded attention module which then inputs to Object prototype extraction module (this part is where I face some challenges ) , it appears it tryign to do similar thing as you do in SafeCountMultiblock layer of model

It is better to contact the authors for their code.