DDU icon indicating copy to clipboard operation
DDU copied to clipboard

Feature Densities Always Zero

Open JacobOaks opened this issue 3 years ago • 2 comments

Hello,

I'm having an issue with the active learning script. I'm running:

CUDA_VISIBLE_DEVICES=7 python active_learning_script.py --seed 1 --model resnet18 -sn -mod --al-type gmm

and I have set a breakpoint right before acquisition in active_learning_script.py here:

Screen Shot 2022-02-04 at 1 12 13 PM

whenever I inspect the result of compute_density(logits, class_prob), density for all instances is always zero like so:

Screen Shot 2022-02-04 at 1 18 38 PM

Upon digging deeper, I noticed that the feature Gaussians always assign extremely low probability to every class, leading to all of these zeros. This happens at every acquisition step for me, regardless of the dataset size and seed (I've tried 2 seeds). I've even tried programmatically setting up a check to tell me if the densities are ever not zero for entire training runs, and it never happens. Thus, for me, the following call to torch.topk amounts to random selection at every acquisition step. I'm wondering if this is an issue that you've experienced.

Thanks, Jacob

JacobOaks avatar Feb 04 '22 18:02 JacobOaks

Hello,

Just checking if you've had a chance to look at this given that it has potentially quite big implications in terms of the validity of the associated paper.

Jacob

JacobOaks avatar Feb 25 '22 18:02 JacobOaks

I have met the same problem!

LinyeLi60 avatar Sep 07 '23 06:09 LinyeLi60