is it possible for you to release the retrain log file?
aptsunny opened this issue · 4 comments
hi, @MingLin-home , In paper and official code, retrain the searched model by the feature loss cost too much resource, the 1.2ms latency model is very difficult to train completely end to end. so whether the series model‘s log can be released?
Hi aptsunny,
Thank you for the feedback!
I cannot find the log file now. Sorry for the bad news :(
ok, at first, I appreciate you for fast response. By the way, I notice the teacher_arch model from a semi-supervised learning approach. i wonder know why do you use the teacher like "geffnet_tf_efficientnet_b3_ns" to distill the student model? it's cause unfair to the baseline model? do you compare the other teacher to distill the searched model before?
thank you in advance.
Hi aptsunny,
Sorry for the late feedback. The teacher network is randomly chosen. We did not test other teacher networks.
Using teacher network is important in NAS. The NAS designed networks are usually deeper and narrower, which means they are more difficult to train. With TS, the improvement of NAS-designed network is often much larger than manually designed networks. For example, TS usually gives 3%~4% accuracy improvement on NAS-designed network, but merely 0.5%~1% on manually designed networks.
Back to ZenNAS. Since ZenNAS always gives us the maximal capacity structures, it is no surprise that with insufficient training data, its generalization error will suffer (due to the curse of dimensionality). Once it is given sufficiently large training data, it can fit much better than other models.
Wish the above answered your concerns! Please let us know if you have more questions!
Best Regards,
Ming Lin