Difference between revisions of "Other Challenges"

From
Jump to: navigation, search
m (BPeat moved page Challenges to Other Challenges without leaving a redirect)
(The Wall - Deep Learning)
Line 18: Line 18:
 
== The Wall - Deep Learning ==
 
== The Wall - Deep Learning ==
 
Deep neural nets are huge and bulky inefficient creatures that allow you to effectively solve a learning problem by getting huge amounts of data and a super computer. They currently trade efficiency for brute force almost every time.
 
Deep neural nets are huge and bulky inefficient creatures that allow you to effectively solve a learning problem by getting huge amounts of data and a super computer. They currently trade efficiency for brute force almost every time.
 +
 +
 +
[https://towardsdatascience.com/recent-advances-for-a-better-understanding-of-deep-learning-part-i-5ce34d1cc914 Towards Theoretical Understanding of Deep Learning | Sanjeev Arora]
 +
 +
* Non Convex Optimization: How can we understand the highly non-convex loss function associated with deep neural networks? Why does stochastic gradient descent even converge?
 +
* Overparametrization and Generalization: In classical statistical theory, generalization depends on the number of parameters but not in deep learning. Why? Can we find another good measure of generalization?
 +
* Role of Depth: How does depth help a neural network to converge? What is the link between depth and generalization?
 +
* Generative Models: Why do Generative Adversarial Networks (GANs) work so well? What theoretical properties could we use to stabilize them or avoid mode collapse?
 +
 +
 +
 +
  
 
<youtube>WTnxE0wjZaM</youtube>
 
<youtube>WTnxE0wjZaM</youtube>

Revision as of 21:02, 16 February 2019

YouTube search... ...Google search

The Wall - Deep Learning

Deep neural nets are huge and bulky inefficient creatures that allow you to effectively solve a learning problem by getting huge amounts of data and a super computer. They currently trade efficiency for brute force almost every time.


Towards Theoretical Understanding of Deep Learning | Sanjeev Arora

  • Non Convex Optimization: How can we understand the highly non-convex loss function associated with deep neural networks? Why does stochastic gradient descent even converge?
  • Overparametrization and Generalization: In classical statistical theory, generalization depends on the number of parameters but not in deep learning. Why? Can we find another good measure of generalization?
  • Role of Depth: How does depth help a neural network to converge? What is the link between depth and generalization?
  • Generative Models: Why do Generative Adversarial Networks (GANs) work so well? What theoretical properties could we use to stabilize them or avoid mode collapse?



The Expert