Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the paper #42

Open
ycjing opened this issue Aug 21, 2020 · 2 comments
Open

About the paper #42

ycjing opened this issue Aug 21, 2020 · 2 comments

Comments

@ycjing
Copy link

ycjing commented Aug 21, 2020

Hi @mrastegari

Thank you for the nice paper and code! The work is really impressive. I have a question about the paper. As mentioned in other issues #11 , Torch does not support bit operations. But in the paper, there is a statement: "This results in 58× faster convolutional operations (in terms of number of the high precision operations) and 32× memory savings". I would appreciate it if you could explain the way to compute these values (i.e., 58x and 32x). Thank you! Wish you all the best!

Best,
Yongcheng

@resurgo97
Copy link

The inference speed is increased on particular accelerators that support bit-wise operations.
Of course it doesn't make it fast to train/inference the model on GPUs.

@ycjing
Copy link
Author

ycjing commented Feb 22, 2021

Hi @resurgo97

Thank you for the response! I appreciate it. Yes, I understand that the improvement in speed can be only achieved on devices that support bit-wise operations. But I wonder how to derive the theoretical improvement in speed (e.g., 58x and 32x). Thank you!

Best,
Yongcheng

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants