aboutsummaryrefslogtreecommitdiff
path: root/microposts/neural-nets-regularization.md
diff options
context:
space:
mode:
Diffstat (limited to 'microposts/neural-nets-regularization.md')
-rw-r--r--microposts/neural-nets-regularization.md8
1 files changed, 0 insertions, 8 deletions
diff --git a/microposts/neural-nets-regularization.md b/microposts/neural-nets-regularization.md
deleted file mode 100644
index 9f2866d..0000000
--- a/microposts/neural-nets-regularization.md
+++ /dev/null
@@ -1,8 +0,0 @@
----
-date: 2018-05-08
----
-> no-one has yet developed an entirely convincing theoretical explanation for why regularization helps networks generalize. Indeed, researchers continue to write papers where they try different approaches to regularization, compare them to see which works better, and attempt to understand why different approaches work better or worse. And so you can view regularization as something of a kludge. While it often helps, we don't have an entirely satisfactory systematic understanding of what's going on, merely incomplete heuristics and rules of thumb.
->
-> There's a deeper set of issues here, issues which go to the heart of science. It's the question of how we generalize. Regularization may give us a computational magic wand that helps our networks generalize better, but it doesn't give us a principled understanding of how generalization works, nor of what the best approach is.
-
-Michael Nielsen, [Neural networks and deep learning](http://neuralnetworksanddeeplearning.com/chap3.html#why_does_regularization_help_reduce_overfitting)