Difference between revisions of "Latent Dirichlet Allocation (LDA)"

From
Jump to: navigation, search
m
m
Line 18: Line 18:
  
  
In [[Natural Language Processing (NLP)]], Latent Dirichlet Allocation (LDA) is a [[Generative AI|generative]] statistical model that allows sets of observations to be explained by unobserved groups that explain why some parts of the data are similar. For example, if observations are words collected into documents, it posits that each document is a mixture of a small number of topics and that each word's presence is attributable to one of the document's topics. LDA is an example of [[Topic Model/Mapping]].
+
In [[Natural Language Processing (NLP)]], [[Latent]] Dirichlet Allocation (LDA) is a [[Generative AI|generative]] statistical model that allows sets of observations to be explained by unobserved groups that explain why some parts of the data are similar. For example, if observations are words collected into documents, it posits that each document is a mixture of a small number of topics and that each word's presence is attributable to one of the document's topics. LDA is an example of [[Topic Model/Mapping]].
  
  

Revision as of 08:33, 16 September 2023

Youtube search... ...Google search


In Natural Language Processing (NLP), Latent Dirichlet Allocation (LDA) is a generative statistical model that allows sets of observations to be explained by unobserved groups that explain why some parts of the data are similar. For example, if observations are words collected into documents, it posits that each document is a mixture of a small number of topics and that each word's presence is attributable to one of the document's topics. LDA is an example of Topic Model/Mapping.