Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

next_layer = curr_layer + 2 during deconv, but it is + 1 in the paper #30

Open
gxlcliqi opened this issue Dec 13, 2017 · 1 comment
Open

Comments

@gxlcliqi
Copy link

Hi,Just notice from the code that sth different between the code and the paper, is there any intend behind that ?

  1. when deconv the next_layer = curr_layer +2 in the code, it means the conv4_1 is deconv by conv2_1 instead of conv3_1, but according to the paper next_layer should be equals curr_layer +1

  2. because of the above implementation, the layer 1 ("conv4_1") is not computed by deconv.

@rozentill
Copy link
Member

rozentill commented Jan 6, 2018

@gxlcliqi
Hi, thanks for your interest on our project.

For deconvolution, the aim is to deconvolve feature map form layer L to the layer L-1, but what we actually do is to deconvolve it to layer L-2 and then forward it to layer L-1. This is because a target feature map in layer L corresponds to many solutions in layer L-1. By incorporating part of the network (form L-2 to L-1 ) into the optimization, we can constrain the feature map we solved is with the same distribution as real network output of layer L-1.

Another way to constrain the solution is to use LBFGS-B and set proper upper and lower bounds. In this way, we can directly deconvolve it to layer L-1.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants