Global AI and Data Science

 View Only
  • 1.  Hyperparameter Tuning in Deep learning

    Posted Wed October 12, 2022 04:54 PM
    Is ReLu the only option as a better optimizer for a neural network to give better performance? And what could be the best possible way to deal with Gradient Vanishing problems in Neural networks?

    ------------------------------
    Pavan Saish Naru
    ------------------------------

    #GlobalAIandDataScience
    #GlobalDataScience


  • 2.  RE: Hyperparameter Tuning in Deep learning

    Posted Thu October 13, 2022 06:20 AM
    Hi Pavan, Im studying for the ML Certification, so not an expert on the subject but at least familiar with it.
    For the dying neurons problem during the learning phase we have methods like Leaky RELU, which allows those previously dead neurons on other methods (RELU, TanH, Sigmoid...) to "leak" part of its weights to the backpropagation calculation. It basically allows you to define a negative slope for your activator so that these nodes contribute also to the calculation. Another method is the Maxout, which is supposed to be superior to Leaky RELY but it doubles the amount of parameters used in the calculation of the gradients.

    As a personal recomendation, one of the best books Ive found on ML is from one of our IBM colleagues, Charu Aggarwal : "Neural Networks and Deep Learning" , if you really want to understand the most complex topics, you can find them here.

    ------------------------------
    Daniel Lopez Sainz
    ------------------------------



  • 3.  RE: Hyperparameter Tuning in Deep learning

    Posted Thu October 13, 2022 07:33 AM

    Hello

    ReLu is not the only option. However, due to way it functions it is one of the fastest activation function. Its good to experiment with other activation functions too, though there is no guarantee of improved model performance on account of activation function alone.






  • 4.  RE: Hyperparameter Tuning in Deep learning

    Posted Thu October 13, 2022 07:36 AM

    To my knowledge, using Leaky ReLu will help reduce impact of vanishing gradient and also reduce impact of dead neurons as in case of ReLu, wherein the negative values become 0.






  • 5.  RE: Hyperparameter Tuning in Deep learning

    Posted Tue October 10, 2023 10:20 AM

    Can you use any of the ReLU activation functions in SPSS 29?  All I see is Sigmoid and Htang.  Can I code in the other functions?  I appreciate your consideration.



    ------------------------------
    Hubert Setzler
    ------------------------------