zlacker

[parent] [thread] 7 comments
1. elgfar+(OP)[view] [source] 2020-04-26 23:00:03
The mathematics of training a neural network. I understand how they work once trained, but that you can train them almost seems too good to be true.
replies(4): >>supern+k3 >>plun9+P4 >>galima+vl >>cambal+Zs
2. supern+k3[view] [source] 2020-04-26 23:32:29
>>elgfar+(OP)
Also, for me, how all these advances in ML/computing are alleged to be on the horizon, when I hear that A/D/C(NN)s are actually so slow in learning. How can something that has to be trained with 1M of <xyz> be smart? What's the next thing?
replies(1): >>OkayPh+601
3. plun9+P4[view] [source] 2020-04-26 23:45:37
>>elgfar+(OP)
You likely understand minimizing a continuously differentiable function. Now, you are minimizing a continuously differentiable error function (which measures the difference between the output of your hypothesis function and actual data), with respect to adjustable weights and biases that determine the value for the neurons going from one layer to the next. The complexity is in that the hypothesis function is a composition of many functions due to the layering, and there usually are a large number of neurons. However, you are basically doing the same thing many times.
4. galima+vl[view] [source] 2020-04-27 02:45:11
>>elgfar+(OP)
A partial answer is that big fully connected neural networks _are_ pretty much untrainable. Neural networks only became successful once programmers started constraining the space they were optimizing over pretty radically (like requiring convolutional layers if you know you are trying to detect something local).
5. cambal+Zs[view] [source] 2020-04-27 04:25:36
>>elgfar+(OP)
On the contrary, it is pretty simple, it is just the same process of "a little bit to the left, a little bit to the right" when you are trying to hang a poster.
replies(1): >>OkayPh+XZ
◧◩
6. OkayPh+XZ[view] [source] [discussion] 2020-04-27 11:50:21
>>cambal+Zs
On the contrary, this is a problem that goes "complicated-easy-complicated" the more you think about it. You're at stage two, so here's the mindfuck: Why aren't we constantly stuck at local minima? Surely these problems we're throwing at ANNs aren't all convex.
replies(1): >>cambal+sM1
◧◩
7. OkayPh+601[view] [source] [discussion] 2020-04-27 11:52:00
>>supern+k3
You and I probably took 4 or 5 years to recognize the alphabet. Cut the machines some slack.
◧◩◪
8. cambal+sM1[view] [source] [discussion] 2020-04-27 17:30:53
>>OkayPh+XZ
How do you KNOW you are not a local minima? You are putting the cart before the horse.
[go to top]