About 50 results
Open links in new tab
  1. What is regularization in plain english? - Cross Validated

    Is regularization really ever used to reduce underfitting? In my experience, regularization is applied on a complex/sensitive model to reduce complexity/sensitvity, but never on a simple/insensitive model to …

  2. What are Regularities and Regularization? - Cross Validated

    Is regularization a way to ensure regularity? i.e. capturing regularities? Why do ensembling methods like dropout, normalization methods all claim to be doing regularization?

  3. L1 & L2 double role in Regularization and Cost functions?

    Mar 19, 2023 · Regularization - penalty for the cost function, L1 as Lasso & L2 as Ridge Cost/Loss Function - L1 as MAE (Mean Absolute Error) and L2 as MSE (Mean Square Error) Are [1] and [2] the …

  4. neural networks - L2 Regularization Constant - Cross Validated

    Dec 3, 2017 · When implementing a neural net (or other learning algorithm) often we want to regularize our parameters $\\theta_i$ via L2 regularization. We do this usually by adding a regularization term …

  5. Why is the L2 regularization equivalent to Gaussian prior?

    Dec 13, 2019 · I keep reading this and intuitively I can see this but how does one go from L2 regularization to saying that this is a Gaussian Prior analytically? Same goes for saying L1 is …

  6. When will L1 regularization work better than L2 and vice versa?

    Nov 29, 2015 · Note: I know that L1 has feature selection property. I am trying to understand which one to choose when feature selection is completely irrelevant. How to decide which regularization (L1 or …

  7. what does regularization mean in xgboost (tree)

    Feb 17, 2019 · In xgboost (xgbtree), gamma is the tunning parameter to control the regularization. I understand what regularization means in xgblinear and logistic regression, but in the context of tree …

  8. Why do we only see $L_1$ and $L_2$ regularization but not other norms?

    Mar 27, 2017 · The intuition behind regularization is that I have some vector, and I would like that vector to be "small" in some sense. How do you describe a vector's size? Well, you have choices: Do you …

  9. How is adding noise to training data equivalent to regularization?

    Oct 18, 2021 · I've noticed that some people argue that adding noise to training data equivalent to regularizing our predictor parameters. How is this the case? Some of the examples listed on SE …

  10. How to explain dropout regularization in simple terms?

    Oct 21, 2016 · If you have a half page to explain dropout, how would you proceed? Which is the rationale behind this technique?