Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

BWN/XNOR SqueezeNet training #13

Open
AGhiuta opened this issue Oct 10, 2016 · 1 comment
Open

BWN/XNOR SqueezeNet training #13

AGhiuta opened this issue Oct 10, 2016 · 1 comment

Comments

@AGhiuta
Copy link

AGhiuta commented Oct 10, 2016

Hi,

I've been trying to train SqueezeNet in both configurations(bwn and xnor), but I can't get past 31% (24% respectively) top-1 accuracy (I was expecting accuracies similar to alexnet). I tried something similar to the GoogLenet variant depicted in the paper (I replaced the expand layers with straightforward convolutions with kernel sizes of 3x3, so there is no branching).

Have you tried to train this model? If positive, can you, please, tell me how did you do it?

Thank you,
Alex

ping @mrastegari

@Rahim16
Copy link

Rahim16 commented Mar 1, 2017

Hi,

I think the problem you are facing arises from that SqueezeNet can achieve 50x reduction in model size compared to AlexNet because it[AlexNet] has a lot of redundancy and also SqueezeNet's Fire modules were fine-tuned to achieve that level of accuracy with float weights. Binarization itself however results in a reduction of redundancy in AlexNet, thus further applying filter size reduction as proposed by SqueezeNet becomes too destructive.

BR,
Rahim

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants