Difference between revisions of "Neural Architecture"

From
Jump to: navigation, search
m
m
Line 15: Line 15:
 
* [[Evolutionary Computation / Genetic Algorithms]]
 
* [[Evolutionary Computation / Genetic Algorithms]]
 
* [[Algorithm Administration#Hyperparameter|Hyperparameter]]s Optimization
 
* [[Algorithm Administration#Hyperparameter|Hyperparameter]]s Optimization
 +
* [[Model Search]]
  
 
= <span id="Neural Architecture Search (NAS)"></span>Neural Architecture Search (NAS) =
 
= <span id="Neural Architecture Search (NAS)"></span>Neural Architecture Search (NAS) =

Revision as of 21:34, 6 March 2022

YouTube search... ...Google search

Neural Architecture Search (NAS)

YouTube search... ...Google search

An alternative to manual design is “neural architecture search” (NAS), a series of machine learning techniques that can help discover optimal neural networks for a given problem. Neural architecture search is a big area of research and holds a lot of promise for future applications of deep learning.

Search spaces for deep learning

deep.jpg

Various approaches to Neural Architecture Search (NAS) have designed networks that are on par or even outperform hand-designed architectures. Methods for NAS can be categorized according to the search space, search strategy and performance estimation strategy used:

  • The search space defines which type of ANN can be designed and optimized in principle.
  • The search strategy defines which strategy is used to find optimal ANN's within the search space.
  • Obtaining the performance of an ANN is costly as this requires training the ANN first. Therefore, performance estimation strategies are used obtain less costly estimates of a model's performance. Neural Architecture Search | Wikipedia

Differentiable Neural Computer (DNC)

[YouTube search...]