Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[D] Why is L2 preferred over L1 Regularization?

I understand L1 regularization induces sparsity, and is thus, good for cases when it’s required.

But In normal use cases, what are the benefits of using L2 over L1? If it’s just that weights should be smaller, then why can’t we use L4 for example?

I’ve seen mentions of L2 capturing energy, Euclidean distance and being rotation invariant. Could one explain these more explicitly as to how this happens?

submitted by /u/tshrjn
[link] [comments]