Transformer

From
Revision as of 10:42, 9 January 2019 by BPeat (talk | contribs)
Jump to: navigation, search

YouTube search... ...Google search


“Attend” to specific parts of the input (an image or text) in sequence, one after another. By relying on a sequence of glances, they capture (visual) structure, can be contrasted with other (machine vision) techniques that process a whole input e.g. image in a single, forward pass.


attention_mechanism.png attention_model.png