I met this equation in Alex paper on local normalization of the answer:
As we see above, the power rises after calculating the sum, multiply it by alpha and then add after adding k.
However, I see that in the TensorFlow documentation it appears as
sqr_sum[a, b, c, d] =
sum(input[a, b, c, d - depth_radius : d + depth_radius + 1] ** 2)
output = input / (bias + alpha * sqr_sum ** beta)
where beta rises only for the amount.
Why is there a discrepancy?
Also, when I looked at the TensorFlow code itself, I saw this:
output[b, r, c, d] /= (
np.power(bias + alpha * np.sum(patch * patch), beta))
which looks right?
I am a bit confused. Can someone correct me please?
source
share