-
Notifications
You must be signed in to change notification settings - Fork 140
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the learning rate setting of p_conv and m_conv #7
Comments
You're right! |
Have you solved this problem now? |
@dontLoveBugs Hello, can you review my issue ? I think the bilinear kernel is wrong |
'tuple' object can not be modified. Your code just get an generator. |
I have searched online, the grad of output can not be modified, if you want modify the grad of input, you need to return the modified grad of input , like : you can try it. My question is : |
@XinZhangNLPR the you is becuse the backforward_hook expected tuple, not 'generator'
Your suggestion still return a generator not a tuple |
it seems this bug has not fixed yet |
You set the gradient of p_conv and m_conv to 0.1 times the other layers, but I find the gradient has no change after backward.
I use the following code to test.
The gradient of p_conv is same with the grad_input, but I think the gradient of p_conv is 0.1 times the gradient of the grad_input. Am I wrong?
The text was updated successfully, but these errors were encountered: