-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Did you reimplement for object detection? #5
Comments
No related plan. In my practice, the effectiveness of distillation depends very much on the data set used. by the way, I also recommend you to try another distillation algorithm - https://github.com/ZJCV/overhaul |
Yup. Recently I have tried many distilltion methods which work on COCO well, but for widerface I can't reproduce a better student. it's very sensitive with the learning rate and other hyper parameters. It would be great if you can share your experience on this. overhaul is on going. |
What scene are you working in? Beside distillation, i also recommend you to try network pruning, which works for me. You can check another two repositories that i did for network pruning. |
Well, I have also tried network slimming or other pruning methods(https://github.com/microsoft/nni) before, but the result is the same as https://arxiv.org/abs/1810.05270. So I am using anytime network to do the pruning now, it's much easier to implement. but I also want to migrate distilling after pruning a network structure. I have experimented KD like overhaul and Knowledge Review on object detection, but I found out they are not stable when training, we may have loss divergence in the middle of the training. other methods like https://github.com/yzd-v/FGD, which is specific for object detection but very hard to tune on my own dataset. |
Hi,
Did you reimplement for object detection? I have tried ReviewKD for my own dataset and my own model, but found out it's not good.
The text was updated successfully, but these errors were encountered: