-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Detail of BinaryDenseNet or BinaryResNet18E #8
Comments
Hi, I agree the framework definitely could benefit from more tutorials and documentation. However it would be best if these hit the "sweet spot" which cover what you (or anyone else who stumbles upon this) is interested in. In the title you name the details of the networks, but in your comment you mention training in general? Which resources have you used/found already? Did you manage to build the framework on your machine? In the meantime, here are a few links, that might help you get started, depending on what you are looking for:
Please let me know if this already helps, or whether you would like additional information. |
Thanks very much for your patience and detailed answer. I have spent days to reproduce your work using pytorch and I already gets 56.3% using your hyperparameter in wiki which indicated your solid work. And I have found your supplementary meterial in arxiv.com and found detailed log of BinaryResNetE of 58.1% and I will reproduce this results too. Thanks again for your detailed log and detailed supplementary material and it really helps us a lot~ |
We also started trying out BinaryDenseNets and the first results seam promising. @LaVieEnRoseSMZ To which supplementary materials are you referring to? It doesn't seam like they are included in the arXiv version: https://arxiv.org/pdf/1906.08637.pdf |
I am referring to the url in the comments of arxiv paper supplementary material |
Thanks. That's helpful 👍 |
I still have one more problem in reading the code of Binarylayer I can not find the defination and implementation of "det_sign" which is used to quantize activation and weight. Can you please show me the url of this part of code? Thanks a lot in advance~ |
As described in Overview of changes, you can find the parts of the code for |
The combination of BMXNet-v2/src/operator/contrib/gradient_cancel-inl.h Lines 97 to 112 in d0aaf81
|
Hi~I am reproducing BinaryDenseNet in the paper. When I go through the code, I find three version of densenet called densenet, densenet_x and densenet_y, what is the exact version used in you final experiments? And I couldn't find DenseNet28 anywhere, would you mind show me the code? Thanks a lot in advance~ |
We use "densenet.py" for all the experiments. The networks like BinaryDenseNet28/37 that have been created by using the densenet specific configurations (init feature number, reduction rate, growth rate) described in the supplementary material published with our paper: https://owncloud.hpi.de/s/1jrAUnqRAfg0TXH |
Thanks for sharing all the details and the supplementary materials. We were able to reproduce your experiments (which unfortunately isn't the case with many other papers) 👍 @LaVieEnRoseSMZ If you are looking for an reimplementation of this paper using Larq (a Keras and TensorFlow based BNN library), you can also checkout the pretrained models and training code at https://larq.dev/models/ and https://github.com/larq/zoo. |
Hi, recently I read your newly realeased paper "Back to Simplicity: How to Train Accurate BNNs from Scratch?" It is a quite good paper and inspires me a lot.
However, I am a little confused about the implementation in this paper. I am not familiar with the code structure of MXNet, Could you please write a more detailed readme or a tutorial or anything similar which could explain the code and the training details?
Thanks a lot in advance~
The text was updated successfully, but these errors were encountered: