Review Request: Muellerklein#20
Review Request: Muellerklein#20FlorianMuellerklein wants to merge 9 commits intoReScience:masterfrom
Conversation
|
AUTHOR There are some formatting issues that I will resolve if this gets accepted. |
|
EDITOR-IN-CHIEF @emmanuelle Any update on this ? |
|
@emmanuelle asked me to review this submission. I need to reinstall my GTX 980Ti to be able to run the code but hopefully I will do it in the coming week. Meanwhile @FlorianMuellerklein is there any change you want to include in the submission? What are the formatting changes you plan? |
|
@ogrisel I had some issues with the image sizes and some captions in the pdf that I need to fix. But, after submitting this pull request someone commented on my github repo that I might have a mistake in the projection shortcut. They said that the shortcut should take the preactivation as it's input. I think they may be correct, should I update my pull request or wait for reviewer input before changing anything? I linked to the issue on my repo below. |
|
@FlorianMuellerklein You can push your changes, they should be tracked here. For the formatting issues in the paper, I will handle it in the edition stage if accepted. @ogrisel Thank you ! |
|
Please @FlorianMuellerklein feel free to push the changes. If you launch a series of new runs please report the new numbers in a new column in your tables so that we can track the impact of that change on the final validation error of the models. |
|
@ogrisel I pushed the changes so everything should be ready to go. I won't have time to check the new numbers for a little while (time to get some more GPUs) but I definitely will. |
|
It seems that the submission misses a file: |
|
@ogrisel That line should be saving the pre-pixel mean to be used to preprocess the testing data later on. I don't understand why it would give an error. |
|
I ok I was in the wrong folder. Edit: Actually no, when I run the script from the top level folder I get this other error: I believe the expectations wrt the folder structure have been broken when converting from the original repo to this submission PR. |
|
Could you please extend the readme to to give the command line parameters to reproduce each of the score reported in the table. I guess those are:
but it would be great if you could confirm this. Also for the sake of reproducibility I think you should include the version number of the libraries you used (and the sha1 digest of the git repo for lasagne). |
|
@ogrisel Ok I'll update the README with the command lines and versions as soon as I can. The depth is actually the multiplier for how many residual blocks to insert between each downsampling layer. So for ResNet-110 and ResNet-164 depth=18 and for the wide residual network depth=2.
|
|
Thanks! |
|
Ok I trained the "wide 2 4" configuration and could reach 0.957476 final validation accuracy. However the |
|
@ogrisel Oh no, I'm really sorry about that. I'll add that to my next update. EDIT: Updated to include python package versions, commands to reproduce. train_nn.py now creates the necessary folders if they don't already exist. |
|
I trained again a "wide 2 4" model and I got: So a 4.86% test error. This is significantly better than what you write in your report. Maybe because of d42627a? I don't have the plot because of the matplotlib backend issue that I fixed in the mean time. You should probably store the train / val losses data into a file to decouple training from convergence monitoring but this is not a big deal. Here is the text log of that run if you are interested. Also it seems that 2 runs don't yield the same results despite the numpy rng seeding at the top of the script. I suspect that the init is done by the theano rng which is probably no seeded. |
|
Another missing folder for the script to work by default: Edit I had launched the run prior to your last change in 0134558. This specific issue should be fixed now. Please ignore this comment. |
|
Here is the result of another run (this time ResNet-110): So the test error for this run of ResNet-110 is 6.15% which is slightly better than the reported 6.37% in the Identity Mappings paper (median of 5 runs). |
|
@ogrisel Wow that is interesting that the wide ResNet did so much better, how close were the two runs with the wide net? Your ResNet-110 is after my latest commit as well? As soon as I get some spare GPU time I will re-run these and update everything on my end. |
|
Dear @emmanuelle @ogrisel @ozancaglayan, Is there anything we can do to have this review to move forward ? |
|
Dear @emmanuelle @ogrisel @ozancaglayan, Any update on this review ? |
|
Sorry for the delay, I am really busy these days. Code
Article
|
|
@FlorianMuellerklein Can you make the suggested modification ? |
|
Should I wait for the modifications to decide? |
|
No, but you can condition your decision (e.g. provided the author makes the suggested changes...) |
|
@rougier Sorry, I've been pretty busy lately as well. I'll have these edits soon. |
|
OK, provided the author makes the suggested changes, my decision is to accept the paper. |
|
@ogrisel Can you tell you decision about the paper or do you need more time ? |
|
Dear @FlorianMuellerklein @ogrisel @emmanuelle @ozancaglayan, This submission is now 6 months old and we need to try to move things forwards. |
|
I just pushed an update with most of the changes that Ozan requested. |
|
@FlorianMuellerklein Thanks for the update! |
|
@ReScience/reviewers We need some new reviewers here if anyone interested! |
|
As I mentioned before, I accept the paper. |
|
@ozancaglayan Sorry, I didn't see it. |
|
🛎 @grisel Can you tell me if you're able to handle this review or if we need to look for another reviewer ? No problem if you can't but I need to know to try to move this review forward. |
|
Sorry for the lack of reply. I will try to find some time next week do review the updated paper / code. |
|
It seems that this uses an unreleased version of Lasagne (which requires a development version of Theano as well). It would be great if the precise Git hashes used to run this could be included in the README. I applaud your including the precise version of numpy and Theano, but you left out sklearn and matplotlib as well. The |
|
@FlorianMuellerklein Are you still following this submission ? We might need to assign a new reviewer to complete the process but first I wanted to know if you're still interested. |
|
If @FlorianMuellerklein is not available, I can jump in - but not sure if it's ideal to have an additional pair of eyes at this point in the process. |
|
Thanks @FedericoV. @FlorianMuellerklein is the author of the replication. @ozancaglayan alread accepted the paper and @ogrisel has not responded yet. At this point, if you can review it, that would be great. |
Muellerklein
Dear @ReScience/editors,
I request a review for the reproduction of the following paper:
and one result from
I believe the original results have been faithfully reproduced as explained in the accompanying article.
Repository lives at https://github.com/FlorianMuellerklein/ReScience-submission/tree/muellerklein
EDITOR
03 July 2016)20 July 2016)24 August 2016)15 March 2017