Difference between revisions of "PRIMO.ai"

From
Jump to: navigation, search
(Classification ...predict categories)
(Learning Techniques)
(14 intermediate revisions by the same user not shown)
Line 30: Line 30:
 
* [[Google AIY Projects Program]]  - Do-it-yourself artificial intelligence
 
* [[Google AIY Projects Program]]  - Do-it-yourself artificial intelligence
 
* [http://www.nvidia.com/en-us/research/ai-playground/ NVIDIA Playground]
 
* [http://www.nvidia.com/en-us/research/ai-playground/ NVIDIA Playground]
 +
 
* [[Competitions]]
 
* [[Competitions]]
Try GPT-2...
+
<i>Try GPT-2...</i>
 
* [http://talktotransformer.com/ Talk to Transformer] - completes your text. | [http://adamdking.com/ Adam D King], [http://huggingface.co/ Hugging Face] and [http://openai.com/ OpenAI]
 
* [http://talktotransformer.com/ Talk to Transformer] - completes your text. | [http://adamdking.com/ Adam D King], [http://huggingface.co/ Hugging Face] and [http://openai.com/ OpenAI]
 
* [http://colab.research.google.com/github/nickwalton/AIDungeon/blob/master/AIDungeon_2.ipynb AI Dungeon 2] AI generated text adventure
 
* [http://colab.research.google.com/github/nickwalton/AIDungeon/blob/master/AIDungeon_2.ipynb AI Dungeon 2] AI generated text adventure
 +
 +
<i>.. more Natural Language Processing (NLP) fun...</i>
 +
* [http://corenlp.run/ CoreNLP - see NLP parsing techniques by pasting your text | Stanford]
 +
* [http://nlp.stanford.edu:8080/sentiment/rntnDemo.html Sentiment Treebank Analysis Demo]
  
 
=== How to... ===
 
=== How to... ===
Line 47: Line 52:
 
= Information Analysis =
 
= Information Analysis =
 
* [[Framing Context]]
 
* [[Framing Context]]
* [[Datasets]]
+
* [[Datasets]] & [[Benchmarks]]
 
* [[Imbalanced Data]]
 
* [[Imbalanced Data]]
 
* [[Data Preprocessing]]
 
* [[Data Preprocessing]]
Line 106: Line 111:
 
* <span id="Unsupervised"></span>[[Unsupervised]]
 
* <span id="Unsupervised"></span>[[Unsupervised]]
 
** [[Radial Basis Function Network (RBFN)]]
 
** [[Radial Basis Function Network (RBFN)]]
** <span id="Self-Supervised"></span>Self-Supervised
+
** <span id="Self-Supervised"></span>[[Self-Supervised]]
refers to an unsupervised learning problem that is framed as a supervised learning problem in order to apply supervised learning algorithms to solve it. A general example of self-supervised learning algorithms are autoencoders. These are a type of neural network that is used to create a compact or compressed representation of an input sample. They achieve this via a model that has an encoder and a decoder element separated by a bottleneck that represents the internal compact representation of the input.
 
 
*** [[Autoencoder (AE) / Encoder-Decoder]]
 
*** [[Autoencoder (AE) / Encoder-Decoder]]
 
*** [[(Stacked) Denoising Autoencoder (DAE)]]
 
*** [[(Stacked) Denoising Autoencoder (DAE)]]
Line 181: Line 185:
  
 
== <span id="Semi-Supervised"></span>[[Semi-Supervised]] ==
 
== <span id="Semi-Supervised"></span>[[Semi-Supervised]] ==
In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters.
+
In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters. Reference: [[Learning Techniques]]
 
* [[Semi-Supervised Learning with Generative Adversarial Network (SSL-GAN)]]
 
* [[Semi-Supervised Learning with Generative Adversarial Network (SSL-GAN)]]
 
* [[Context-Conditional Generative Adversarial Network (CC-GAN)]]
 
* [[Context-Conditional Generative Adversarial Network (CC-GAN)]]
Line 187: Line 191:
 
== <span id="Natural Language"></span>Natural Language  ==
 
== <span id="Natural Language"></span>Natural Language  ==
  
* [[Natural Language Processing (NLP)]] involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis  
+
* [[Natural Language Processing (NLP)]] involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis
** Current State of the Art:
 
*** [[Attention Mechanism/Model - Transformer Model]]
 
**** [[Generative Pre-trained Transformer-2 (GPT-2)]] ..[http://talktotransformer.com/ Talk To Transformer]
 
**** [[Bidirectional Encoder Representations from Transformers (BERT)]]
 
**** [[XLNet]] extends [[Transformer-XL]]
 
*** [[(Deep) Convolutional Neural Network (DCNN/CNN)]]
 
*** [[Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), and Recurrent Neural Network (RNN)]]
 
** Methods:
 
*** [[Natural Language Processing (NLP)#Text Preprocessing |Text Preprocessing]]
 
**** [[Natural Language Processing (NLP)#Text Regular Expressions (Regex) |Regular Expressions (Regex)]]
 
**** [[Natural Language Processing (NLP)#Soundex |Soundex]]
 
**** [[Natural Language Processing (NLP)#Tokenization / Sentence Splitting |Tokenization / Sentence Splitting]]
 
***** [[Natural Language Processing (NLP)#Word Embeddings |Word Embeddings]]
 
**** [[Natural Language Processing (NLP)#Normalization |Normalization]]
 
***** [[Natural Language Processing (NLP)#Stemming (Morphological Similarity) |Stemming (Morphological Similarity)]]
 
***** [[Natural Language Processing (NLP)#Lemmatization |Lemmatization]]
 
**** [[Natural Language Processing (NLP)#Similarity |Similarity]]
 
***** [[Natural Language Processing (NLP)#Word Similarity |Word Similarity]]
 
***** [[Natural Language Processing (NLP)#Text Clustering |Text Clustering]]
 
***** [[Natural Language Processing (NLP)#Sentence/Document Similarity |Sentence/Document Similarity]]
 
***** [[Natural Language Processing (NLP)#Text Classification |Text Classification]]
 
***** [[Natural Language Processing (NLP)#Topic Modeling |Topic Modeling]]
 
**** [[Natural Language Processing (NLP)#Whole Word Masking |Whole Word Masking]]
 
**** [[Natural Language Processing (NLP)#Identity Scrubbing |Identity Scrubbing]]
 
**** [[Natural Language Processing (NLP)#Stop Words |Stop Words]]
 
*** [[Natural Language Processing (NLP)#Relating Text |Relating Text]]
 
**** [[Natural Language Processing (NLP)#Part-of-Speech (POS) Tagging |Part-of-Speech (POS) Tagging]]
 
**** [[Natural Language Processing (NLP)#Chunking |Chunking]] - chunks or patterns, e.g. telephone number
 
**** [[Natural Language Processing (NLP)#Chinking |Chinking]] - unwanted chunk removal
 
**** [[Natural Language Processing (NLP)#Named Entity Recognition (NER) |Named Entity Recognition (NER)]]
 
**** [[Natural Language Processing (NLP)#Relation Extraction |Relation Extraction]]
 
**** [[Natural Language Processing (NLP)#Neural Coreference |Neural Coreference]]
 
** [[Natural Language Processing (NLP)#Natural Language Understanding (NLU) |Natural Language Understanding (NLU)]] or Natural Language Interpretation (NLI)
 
*** [[Natural Language Processing (NLP)#Managed Vocabularies |Managed Vocabularies]]
 
**** [[Natural Language Processing (NLP)#Corpora |Corpora]]
 
**** [[Natural Language Processing (NLP)#Ontologies |Ontologies]] and [[Natural Language Processing (NLP)#Taxonomies |Taxonomies]]
 
*** [[Natural Language Processing (NLP)#Natural Language Inference (NLI) and Recognizing Textual Entailment (RTE)|Natural Language Inference (NLI) and Recognizing Textual Entailment (RTE)]]
 
**** [[Natural Language Processing (NLP)#Semantic Role Labeling (SRL) |Semantic Role Labeling (SRL)]]
 
*** [[Natural Language Processing (NLP)#Deep Learning Algorithms |Deep Learning Algorithms]]
 
*** [[Natural Language Processing (NLP)#Capabilities |Capabilities]]
 
**** [[Natural Language Processing (NLP)#Summarization |Summarization]]
 
**** [[Natural Language Processing (NLP)#Sentiment Analysis |Sentiment Analysis]]
 
**** [[Natural Language Processing (NLP)#Wikifier |Wikifier]]
 
*** [[Natural Language Processing (NLP)#Workbench / Pipeline |Workbench / Pipeline]]
 
* [[Natural Language Generation (NLG)]] involves writing/generating complete grammatically correct sentences and paragraphs
 
  
 
== <span id="Reinforcement Learning (RL)"></span>[[Reinforcement Learning (RL)]]  ==
 
== <span id="Reinforcement Learning (RL)"></span>[[Reinforcement Learning (RL)]]  ==
Line 303: Line 262:
 
* [[Lifelong Learning]] - Catastrophic Forgetting Challenge
 
* [[Lifelong Learning]] - Catastrophic Forgetting Challenge
 
* [[Neural Structured Learning (NSL)]]
 
* [[Neural Structured Learning (NSL)]]
 +
* [[Meta-Learning]]
  
 
=== Opportunities & Challenges ===
 
=== Opportunities & Challenges ===

Revision as of 09:59, 23 February 2020

On Saturday April 20, 2024 PRIMO.ai has 736 pages

Getting Started

Overview

Background

AI Breakthroughs

AI Fun

Try GPT-2...

.. more Natural Language Processing (NLP) fun...

How to...

Forward Thinking

Information Analysis

Algorithms

Predict values - Regression

Classification ...predict categories

Recommendation

Clustering - Continuous - Dimensional Reduction

Hierarchical

Convolutional

Deconvolutional

Graph

- includes social networks, sensor networks, the entire Internet, 3D Objects (Point Cloud)

Sequence / Time

Time

Spatialtemporal

Spatial-Temporal Dynamic Network (STDN)

Competitive

Semi-Supervised

In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters. Reference: Learning Techniques

Natural Language

  • Natural Language Processing (NLP) involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis

Reinforcement Learning (RL)

an algorithm receives a delayed reward in the next time step to evaluate its previous action. Therefore based on those decisions, the algorithm will train itself based on the success/error of output. In combination with Neural Networks it is capable of solving more complex tasks. Policy Gradient (PG) methods are a type of reinforcement learning techniques that rely upon optimizing parametrized policies with respect to the expected return (long-term cumulative reward) by gradient descent.

Neuro-Symbolic

the “connectionists” seek to construct artificial neural networks, inspired by biology, to learn about the world, while the “symbolists” seek to build intelligent machines by coding in logical rules and representations of the world. Neuro-Symbolic combines the fruits of group.

Other

Techniques

Methods & Concepts

Learning Techniques

Opportunities & Challenges

Development & Implementation

No Coding

Coding

Libraries & Frameworks

TensorFlow

Tooling

Platforms: Machine Learning as a Service (MLaaS)

... and other leading organizations





If you get a 502 or 503 error please try the webpage again, as your message is visiting the island which the server is located, perhaps deciding to relax in the Sun before returning. Thank you.