Feature Densities Always Zero
JacobOaks opened this issue · 2 comments
Hello,
I'm having an issue with the active learning script. I'm running:
CUDA_VISIBLE_DEVICES=7 python active_learning_script.py --seed 1 --model resnet18 -sn -mod --al-type gmm
and I have set a breakpoint right before acquisition in active_learning_script.py
here:
whenever I inspect the result of compute_density(logits, class_prob)
, density for all instances is always zero like so:
Upon digging deeper, I noticed that the feature Gaussians always assign extremely low probability to every class, leading to all of these zeros. This happens at every acquisition step for me, regardless of the dataset size and seed (I've tried 2 seeds). I've even tried programmatically setting up a check to tell me if the densities are ever not zero for entire training runs, and it never happens. Thus, for me, the following call to torch.topk
amounts to random selection at every acquisition step. I'm wondering if this is an issue that you've experienced.
Thanks,
Jacob
Hello,
Just checking if you've had a chance to look at this given that it has potentially quite big implications in terms of the validity of the associated paper.
Jacob
I have met the same problem!