12-28-2015, 06:26 AM
(12-28-2015, 04:34 AM)Bear Wrote: I was trying to find a solution (or at least a heuristic) for the Stable Initialization Problem, which is actually relatively minor in Artificial Neural Networks these days. Because autoencoding strategies allow for training deep networks, Stable Initialization has turned into more of a "would be nice to have" than a "blocks all progress" sort of problem.
But one of the things I tried turns out to crack the Catastrophic Forgetting Problem, the Exploding and Vanishing Gradient Problems, and automates large-scale Adaptive Learning Rates on the basis of purely local information. And those four things are MAJOR problems.
I know some people who are interested in general artificial intelligences; is there anything you can share about the nature of your insight, publicly or privately?
Thank you for your time,
--
DataPacRat
"Does aₘᵢₙ=2c²/Θ ? I don't know, but wouldn't it be fascinating if it were?"
--
DataPacRat
"Does aₘᵢₙ=2c²/Θ ? I don't know, but wouldn't it be fascinating if it were?"