Difference between revisions of "PRIMO.ai"

From
Jump to: navigation, search
(Advanced Learning)
(41 intermediate revisions by the same user not shown)
Line 4: Line 4:
 
|keywords=artificial, intelligence, machine, learning, models, algorithms, cybersecurity, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS  
 
|keywords=artificial, intelligence, machine, learning, models, algorithms, cybersecurity, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS  
 
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools  
 
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools  
 +
 +
<script async src="https://www.googletagmanager.com/gtag/js?id=UA-101589255-2"></script>
 +
<!-- Global site tag (gtag.js) - Google Analytics -->
 +
<script>
 +
  window.dataLayer = window.dataLayer || [];
 +
  function gtag(){dataLayer.push(arguments);}
 +
  gtag('js', new Date());
 +
 +
  gtag('config', 'UA-101589255-2');
 +
</script>
 
}}
 
}}
 
 
On {{LOCALDAYNAME}} {{LOCALMONTHNAME}} {{LOCALDAY}}, {{LOCALYEAR}} PRIMO.ai has {{NUMBEROFPAGES}} pages  
 
On {{LOCALDAYNAME}} {{LOCALMONTHNAME}} {{LOCALDAY}}, {{LOCALYEAR}} PRIMO.ai has {{NUMBEROFPAGES}} pages  
  
Line 30: Line 39:
 
* [[Google AIY Projects Program]]  - Do-it-yourself artificial intelligence
 
* [[Google AIY Projects Program]]  - Do-it-yourself artificial intelligence
 
* [http://www.nvidia.com/en-us/research/ai-playground/ NVIDIA Playground]
 
* [http://www.nvidia.com/en-us/research/ai-playground/ NVIDIA Playground]
* [http://talktotransformer.com/ Try GPT-2...Talk to Transformer] - completes your text. | [http://adamdking.com/ Adam D King], [http://huggingface.co/ Hugging Face] and [http://openai.com/ OpenAI]
+
 
 
* [[Competitions]]
 
* [[Competitions]]
 +
<i>Try GPT-2...</i>
 +
* [http://talktotransformer.com/ Talk to Transformer] - completes your text. | [http://adamdking.com/ Adam D King], [http://huggingface.co/ Hugging Face] and [http://openai.com/ OpenAI]
 +
* [http://colab.research.google.com/github/nickwalton/AIDungeon/blob/master/AIDungeon_2.ipynb AI Dungeon 2] AI generated text adventure
 +
 +
<i>.. more Natural Language Processing (NLP) fun...</i>
 +
* [http://corenlp.run/ CoreNLP - see NLP parsing techniques by pasting your text | Stanford]
 +
* [http://nlp.stanford.edu:8080/sentiment/rntnDemo.html Sentiment Treebank Analysis Demo]
  
 
=== How to... ===
 
=== How to... ===
Line 45: Line 61:
 
= Information Analysis =
 
= Information Analysis =
 
* [[Framing Context]]
 
* [[Framing Context]]
* [[Datasets]]
+
* [[Datasets]] & [[Benchmarks]]
 
* [[Imbalanced Data]]
 
* [[Imbalanced Data]]
 
* [[Data Preprocessing]]
 
* [[Data Preprocessing]]
Line 64: Line 80:
  
 
= [[Algorithms]] =
 
= [[Algorithms]] =
 
+
* [[Model Zoos]]
 +
* [[Graphical Tools for Modeling AI Components]]
  
 
== Predict values - [[Regression]] ==  
 
== Predict values - [[Regression]] ==  
Line 84: Line 101:
  
 
== Classification [[...predict categories]] ==
 
== Classification [[...predict categories]] ==
* [[Supervised]]
+
* <span id="Supervised"></span>[[Supervised]]
 
** [[Naive Bayes]]
 
** [[Naive Bayes]]
 
** [[K-Nearest Neighbors (KNN)]]
 
** [[K-Nearest Neighbors (KNN)]]
Line 101: Line 118:
 
** [[Apriori, Frequent Pattern (FP) Growth, Association Rules/Analysis]]
 
** [[Apriori, Frequent Pattern (FP) Growth, Association Rules/Analysis]]
 
** [[Markov Model (Chain, Discrete Time, Continuous Time, Hidden)]]
 
** [[Markov Model (Chain, Discrete Time, Continuous Time, Hidden)]]
* [[Unsupervised]]
+
* <span id="Unsupervised"></span>[[Unsupervised]]
 
** [[Radial Basis Function Network (RBFN)]]
 
** [[Radial Basis Function Network (RBFN)]]
** [[Autoencoder (AE) / Encoder-Decoder]]
+
** <span id="Self-Supervised"></span>[[Self-Supervised]]
** [[(Stacked) Denoising Autoencoder (DAE)]]
+
*** [[Autoencoder (AE) / Encoder-Decoder]]
** [[Sparse Autoencoder (SAE)]]
+
*** [[(Stacked) Denoising Autoencoder (DAE)]]
 +
*** [[Sparse Autoencoder (SAE)]]
  
 
== [[Recommendation]] ==
 
== [[Recommendation]] ==
Line 175: Line 193:
 
* [[Quantum Generative Adversarial Learning (QuGAN - QGAN)]]
 
* [[Quantum Generative Adversarial Learning (QuGAN - QGAN)]]
  
== [[Semi-Supervised]] ==
+
== <span id="Semi-Supervised"></span>[[Semi-Supervised]] ==
In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters.
+
In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters. Reference: [[Learning Techniques]]
 
* [[Semi-Supervised Learning with Generative Adversarial Network (SSL-GAN)]]
 
* [[Semi-Supervised Learning with Generative Adversarial Network (SSL-GAN)]]
 
* [[Context-Conditional Generative Adversarial Network (CC-GAN)]]
 
* [[Context-Conditional Generative Adversarial Network (CC-GAN)]]
  
== Natural Language ==
+
== <span id="Natural Language"></span>Natural Language  ==
  
* [[Natural Language Processing (NLP)]] involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis  
+
* [[Natural Language Processing (NLP)]] involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis
** Current State of the Art:
 
*** [[(Deep) Convolutional Neural Network (DCNN/CNN)]]
 
*** [[Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), and Recurrent Neural Network (RNN)]]
 
*** [[Attention Mechanism/Model - Transformer Model]]
 
*** [[Bidirectional Encoder Representations from Transformers (BERT)]]
 
*** [[XLNet]] extends [[Transformer-XL]]
 
** Methods:
 
*** [[Natural Language Processing (NLP)#Text Preprocessing |Text Preprocessing]]
 
**** [[Natural Language Processing (NLP)#Text Regular Expressions (Regex) |Regular Expressions (Regex)]]
 
**** [[Natural Language Processing (NLP)#Soundex |Soundex]]
 
**** [[Natural Language Processing (NLP)#Tokenization / Sentence Splitting |Tokenization / Sentence Splitting]]
 
***** [[Natural Language Processing (NLP)#Word Embeddings |Word Embeddings]]
 
**** [[Natural Language Processing (NLP)#Normalization |Normalization]]
 
***** [[Natural Language Processing (NLP)#Stemming (Morphological Similarity) |Stemming (Morphological Similarity)]]
 
***** [[Natural Language Processing (NLP)#Lemmatization |Lemmatization]]
 
**** [[Natural Language Processing (NLP)#Similarity |Similarity]]
 
***** [[Natural Language Processing (NLP)#Word Similarity |Word Similarity]]
 
***** [[Natural Language Processing (NLP)#Text Clustering |Text Clustering]]
 
***** [[Natural Language Processing (NLP)#Sentence/Document Similarity |Sentence/Document Similarity]]
 
***** [[Natural Language Processing (NLP)#Text Classification |Text Classification]]
 
***** [[Natural Language Processing (NLP)#Topic Modeling |Topic Modeling]]
 
**** [[Natural Language Processing (NLP)#Whole Word Masking |Whole Word Masking]]
 
**** [[Natural Language Processing (NLP)#Identity Scrubbing |Identity Scrubbing]]
 
**** [[Natural Language Processing (NLP)#Stop Words |Stop Words]]
 
*** [[Natural Language Processing (NLP)#Relating Text |Relating Text]]
 
**** [[Natural Language Processing (NLP)#Part-of-Speech (POS) Tagging |Part-of-Speech (POS) Tagging]]
 
**** [[Natural Language Processing (NLP)#Chunking |Chunking]] - chunks or patterns, e.g. telephone number
 
**** [[Natural Language Processing (NLP)#Chinking |Chinking]] - unwanted chunk removal
 
**** [[Natural Language Processing (NLP)#Named Entity Recognition (NER) |Named Entity Recognition (NER)]]
 
**** [[Natural Language Processing (NLP)#Relation Extraction |Relation Extraction]]
 
**** [[Natural Language Processing (NLP)#Neural Coreference |Neural Coreference]]
 
** [[Natural Language Processing (NLP)#Natural Language Understanding (NLU) |Natural Language Understanding (NLU)]] or Natural Language Interpretation (NLI)
 
*** [[Natural Language Processing (NLP)#Managed Vocabularies |Managed Vocabularies]]
 
**** [[Natural Language Processing (NLP)#Corpora |Corpora]]
 
**** [[Natural Language Processing (NLP)#Ontologies |Ontologies]] and [[Natural Language Processing (NLP)#Taxonomies |Taxonomies]]
 
*** [[Natural Language Processing (NLP)#Natural Language Inference (NLI) and Recognizing Textual Entailment (RTE)|Natural Language Inference (NLI) and Recognizing Textual Entailment (RTE)]]
 
**** [[Natural Language Processing (NLP)#Semantic Role Labeling (SRL) |Semantic Role Labeling (SRL)]]
 
*** [[Natural Language Processing (NLP)#Deep Learning Algorithms |Deep Learning Algorithms]]
 
*** [[Natural Language Processing (NLP)#Capabilities |Capabilities]]
 
**** [[Natural Language Processing (NLP)#Summarization |Summarization]]
 
**** [[Natural Language Processing (NLP)#Sentiment Analysis |Sentiment Analysis]]
 
**** [[Natural Language Processing (NLP)#Wikifier |Wikifier]]
 
*** [[Natural Language Processing (NLP)#Workbench / Pipeline |Workbench / Pipeline]]
 
* [[Natural Language Generation (NLG)]] involves writing/generating complete grammatically correct sentences and paragraphs
 
  
== [[Reinforcement Learning (RL)]]  ==
+
== <span id="Reinforcement Learning (RL)"></span>[[Reinforcement Learning (RL)]]  ==
 
an algorithm receives a delayed reward in the next time step to evaluate its previous action. Therefore based on those decisions, the algorithm will train itself based on the success/error of output. In combination with Neural Networks it is capable of solving more complex tasks. [[Policy Gradient (PG)]] methods are a type of reinforcement learning techniques that rely upon optimizing parametrized policies with respect to the expected return (long-term cumulative reward) by [[Gradient Descent Optimization & Challenges |gradient descent]].
 
an algorithm receives a delayed reward in the next time step to evaluate its previous action. Therefore based on those decisions, the algorithm will train itself based on the success/error of output. In combination with Neural Networks it is capable of solving more complex tasks. [[Policy Gradient (PG)]] methods are a type of reinforcement learning techniques that rely upon optimizing parametrized policies with respect to the expected return (long-term cumulative reward) by [[Gradient Descent Optimization & Challenges |gradient descent]].
  
Line 285: Line 259:
 
* [[Local Features]]
 
* [[Local Features]]
  
=== [[Advanced Learning]] ===
+
=== [[Learning Techniques]] ===
 
* [[Text Transfer Learning]]  
 
* [[Text Transfer Learning]]  
 
* [[Image/Video Transfer Learning]]
 
* [[Image/Video Transfer Learning]]
Line 297: Line 271:
 
* [[Lifelong Learning]] - Catastrophic Forgetting Challenge
 
* [[Lifelong Learning]] - Catastrophic Forgetting Challenge
 
* [[Neural Structured Learning (NSL)]]
 
* [[Neural Structured Learning (NSL)]]
 +
* [[Meta-Learning]]
  
 
=== Opportunities & Challenges ===
 
=== Opportunities & Challenges ===
Line 313: Line 288:
 
** [[Messaging & Routing]]  
 
** [[Messaging & Routing]]  
 
** [[Pipeline]]s
 
** [[Pipeline]]s
** [[Federated]]
+
** [[Federated]] Learning
 +
** [[Distributed]] Learning
 
** [[Processing Units - CPU, GPU, APU, TPU, VPU, FPGA, QPU]]
 
** [[Processing Units - CPU, GPU, APU, TPU, VPU, FPGA, QPU]]
 
* [[Integrity Forensics]]
 
* [[Integrity Forensics]]
Line 363: Line 339:
 
* [[Intel]]
 
* [[Intel]]
 
* [[Apple]]
 
* [[Apple]]
 +
* [[IBM]]
 +
 +
=== ... and other leading organizations ===
 +
* [http://allenai.org/ Allen Institute for Artificial Intelligence, or AI2]
 +
* [http://openai.com/ OpenAI]
 +
  
==== [http://machinelearning.apple.com/ Apple] ====
 
* [[Turi]]
 
  
  

Revision as of 21:08, 22 May 2020

On Tuesday April 16, 2024 PRIMO.ai has 736 pages

Getting Started

Overview

Background

AI Breakthroughs

AI Fun

Try GPT-2...

.. more Natural Language Processing (NLP) fun...

How to...

Forward Thinking

Information Analysis

Algorithms

Predict values - Regression

Classification ...predict categories

Recommendation

Clustering - Continuous - Dimensional Reduction

Hierarchical

Convolutional

Deconvolutional

Graph

- includes social networks, sensor networks, the entire Internet, 3D Objects (Point Cloud)

Sequence / Time

Time

Spatialtemporal

Spatial-Temporal Dynamic Network (STDN)

Competitive

Semi-Supervised

In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters. Reference: Learning Techniques

Natural Language

  • Natural Language Processing (NLP) involves speech recognition, (speech) translation, understanding (semantic parsing) complete sentences, understanding synonyms of matching words, and sentiment analysis

Reinforcement Learning (RL)

an algorithm receives a delayed reward in the next time step to evaluate its previous action. Therefore based on those decisions, the algorithm will train itself based on the success/error of output. In combination with Neural Networks it is capable of solving more complex tasks. Policy Gradient (PG) methods are a type of reinforcement learning techniques that rely upon optimizing parametrized policies with respect to the expected return (long-term cumulative reward) by gradient descent.

Neuro-Symbolic

the “connectionists” seek to construct artificial neural networks, inspired by biology, to learn about the world, while the “symbolists” seek to build intelligent machines by coding in logical rules and representations of the world. Neuro-Symbolic combines the fruits of group.

Other

Techniques

Methods & Concepts

Learning Techniques

Opportunities & Challenges

Development & Implementation

No Coding

Coding

Libraries & Frameworks

TensorFlow

Tooling

Platforms: Machine Learning as a Service (MLaaS)

... and other leading organizations





If you get a 502 or 503 error please try the webpage again, as your message is visiting the island which the server is located, perhaps deciding to relax in the Sun before returning. Thank you.