You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Neuron Pruning with L-inf norm is better than with L-2 norm 💡✨
-------------------------------------------
As you know, Pruning is a process to compress the model without decrease the accuracy by neglect unimportant weights or neurons.
To prune a neuron based on weight magnitude you can use the L-2 norm of the neuron’s weights.
in this example, the author use neural network with just 4 hidden layers in Minst dataset and get accuracy around 90%. after pruning around 25% of neurons with L-2 norm, the accuracy became 88%. but with L-inf norm you can prune 40% to have 88% accuracy also. and if you prune 60% of neurons, the accuracy will drop 9% and became 80%.
So, why L-inf norm is better than L-2 norm ?
why don't we use it ? also in Regularization for models ?