About 60,400 results
Open links in new tab
  1. Update equation for gradient descent - Stack Overflow

    Dec 14, 2015 · Recall, gradient descent is based on the Taylor expansion of f (w, x) in the close vicinity of w, and has its purpose---in your context---in repeatedly modifying the weight in small …

  2. gradient descent using python and numpy - Stack Overflow

    Jul 22, 2013 · As you can see I also added the generated regression line and formula that was calculated by excel. You need to take care about the intuition of the regression using gradient …

  3. simultaneously update theta0 and theta1 to calculate gradient …

    Sep 2, 2019 · 5 I am taking the machine learning course from coursera. There is a topic called gradient descent to optimize the cost function. It says to simultaneously update theta0 and …

  4. python - Gradient descent for ridge regression - Stack Overflow

    Jan 26, 2021 · I'm trying to write a code that return the parameters for ridge regression using gradient descent. Ridge regression is defined as Where, L is the loss (or cost) function. w are …

  5. Multivariate Linear Regression using gradient descent

    Dec 13, 2020 · For your purposes you can follow the following formula to randomly initialize the weights using numpy's random.randn where l is a particular layer. This will result in the …

  6. How to update the bias in neural network backpropagation?

    Sep 23, 2010 · If you imagine your network as a traveler trying to reach the bottom of these gradients (i.e. Gradient Descent), then the amount you will change each weight (and bias) by …

  7. What is the difference between Gradient Descent and Newton's …

    82 I understand what Gradient Descent does. Basically it tries to move towards the local optimal solution by slowly moving down the curve. I am trying to understand what is the actual …

  8. neural network - Different delta rules - Stack Overflow

    Dec 18, 2019 · The first formula is a general expression while the second is a rule on how to compute the gradient coefficient in function of the previous gradient. The new weights are …

  9. python - Sklearn LinearRegression () don't require Iterations and ...

    Mar 4, 2024 · LinearRegression() doesn't use gradient descent, which is why it does not have a learning_rate parameter. Instead, it uses a mathematical formula for directly computing the …

  10. Adam Optimizer vs Gradient Descent - Stack Overflow

    Aug 25, 2018 · Its is an adaptive method compared to the gradient descent which maintains a single learning rate for all weight updates and the learning rate does not change. Adam has …