Submitted by AutoModerator t3_z07o4c in MachineLearning
Please post your questions here instead of creating a new thread. Encourage others who create new posts for questions to post here instead!
Thread will stay alive until next one so keep posting after the date in the title.
Thanks to everyone for answering questions in the previous thread!
koiRitwikHai t1_ix93rg6 wrote
Why ReLU performs better than other activation functions when it is neither a differentiable function nor it is zero-centered?