Difference between revisions of "Diffusion"

From
Jump to: navigation, search
m
m
 
(18 intermediate revisions by the same user not shown)
Line 9: Line 9:
  
 
* [[Markov Model (Chain, Discrete Time, Continuous Time, Hidden)]]
 
* [[Markov Model (Chain, Discrete Time, Continuous Time, Hidden)]]
* [https://stablediffusionweb.com/ Stable Diffusion | Stable Diffusion] ... a latent text-to-image diffusion model capable of generating photo-realistic images given any text input
+
* [https://stablediffusionweb.com/ Stable Diffusion | Stable Diffusion] ... a [[latent]] text-to-image diffusion model capable of generating photo-realistic images given any text input
* [https://openai.com/dall-e-2/ DALL-E 2 |] [[OpenAI]]
+
* [[Music#Stable Audio|Stable Audio]]
 +
* [[Video/Image#DALL-E | DALL-E]]
 
** [https://www.assemblyai.com/blog/how-dall-e-2-actually-works/ How DALL-E 2 Actually Works | Ryan O'Connor - AssemblyAI]
 
** [https://www.assemblyai.com/blog/how-dall-e-2-actually-works/ How DALL-E 2 Actually Works | Ryan O'Connor - AssemblyAI]
 +
* [[What is Artificial Intelligence (AI)? | Artificial Intelligence (AI)]] ... [[Generative AI]] ... [[Machine Learning (ML)]] ... [[Deep Learning]] ... [[Neural Network]] ... [[Reinforcement Learning (RL)|Reinforcement]] ... [[Learning Techniques]]
 +
* [[Conversational AI]] ... [[ChatGPT]] | [[OpenAI]] ... [[Bing/Copilot]] | [[Microsoft]] ... [[Gemini]] | [[Google]] ... [[Claude]] | [[Anthropic]] ... [[Perplexity]] ... [[You]] ... [[phind]] ... [[Grok]] | [https://x.ai/ xAI] ... [[Groq]] ... [[Ernie]] | [[Baidu]]
 +
* [https://www.cloudskillsboost.google/course_templates/541 Introduction to Image Generation | Google] ... course introduces diffusion models, a family of machine learning models that recently showed promise in the image generation space.
  
Diffusion Models are generative models, meaning that they are used to generate data similar to the data on which they are trained. Fundamentally, Diffusion Models work by destroying training data through the successive addition of Gaussian noise, and then learning to recover the data by reversing this noising process. [https://www.assemblyai.com/blog/diffusion-models-for-machine-learning-introduction/#:~:text=Diffusion%20Models%20are%20generative%20models%2C%20meaning%20that%20they%20are%20used,by%20reversing%20this%20noising%20process. | Ryan O'Connor - AssmblyAI]
+
Diffusion Models are [[Generative AI|generative]] models, meaning that they are used to generate data similar to the data on which they are trained. Fundamentally, Diffusion Models work by destroying training data through the successive addition of Gaussian noise, and then learning to recover the data by reversing this noising process. [https://www.assemblyai.com/blog/diffusion-models-for-machine-learning-introduction/#:~:text=Diffusion%20Models%20are%20generative%20models%2C%20meaning%20that%20they%20are%20used,by%20reversing%20this%20noising%20process. | Ryan O'Connor - AssmblyAI]
  
+
 
 +
<youtube>zc5NTeJbk-k</youtube>
 
<youtube>yTAMrHVG1ew</youtube>
 
<youtube>yTAMrHVG1ew</youtube>
 
<youtube>HoKDTa5jHvg</youtube>
 
<youtube>HoKDTa5jHvg</youtube>
 
<youtube>a4Yfz2FxXiY</youtube>
 
<youtube>a4Yfz2FxXiY</youtube>
 
<youtube>DHaL56P6f5M</youtube>
 
<youtube>DHaL56P6f5M</youtube>
 +
<youtube>S-Js1srE2CE</youtube>
 +
<youtube>jytmYCC5OtE</youtube>
 +
 +
= Control Diffusion Models =
 +
* [https://arxiv.org/abs/2302.05543 Adding Conditional Control to Text-to-Image Diffusion Models | Lvmin Zhang, Maneesh Agrawala]
 +
* [https://arxiv.org/abs/2302.03668 Hard Prompts Made Easy: Gradient-Based Discrete Optimization for Prompt Tuning and Discovery | Y. Wen, N. Jain, J. Kirchenbauer, M. Goldblum, J. Geiping, and T. Goldstein]
 +
 +
[https://github.com/lllyasviel/ControlNet ControlNet | llyasviel]:  We present a neural network structure, ControlNet, to control pretrained large diffusion models to support additional input conditions. The ControlNet learns task-specific conditions in an end-to-end way, and the learning is robust even when the training dataset is small (< 50k). Moreover, training a ControlNet is as fast as [[fine-tuning]] a diffusion model, and the model can be trained on a personal devices. Alternatively, if powerful computation clusters are available, the model can scale to large amounts (millions to billions) of data. We report that large diffusion models like Stable Diffusion can be augmented with ControlNets to enable conditional inputs like edge maps, segmentation maps, keypoints, etc. This may enrich the methods to control large diffusion models and further facilitate related applications. [https://arxiv.org/abs/2302.05543 Adding Conditional Control to Text-to-Image Diffusion Models | Lvmin Zhang, Maneesh Agrawala]
 +
 +
There exists a very subtle but powerful methodology for deep learning. This simple idea is to use a neural network to extract the “specification” from an example. Then use another neural network to generate examples from the extracted example. This is a universal idea.  [https://medium.com/intuitionmachine/deep-learning-modularity-and-language-models-bd726c5e3b58 Deep Learning Modularity and Language Models | Carlos E. Perez - Medium]
 +
 +
https://github.com/lllyasviel/ControlNet/raw/main/github_page/he.png

Latest revision as of 06:44, 11 May 2024

YouTube search... ...Google search

Diffusion Models are generative models, meaning that they are used to generate data similar to the data on which they are trained. Fundamentally, Diffusion Models work by destroying training data through the successive addition of Gaussian noise, and then learning to recover the data by reversing this noising process. | Ryan O'Connor - AssmblyAI


Control Diffusion Models

ControlNet | llyasviel: We present a neural network structure, ControlNet, to control pretrained large diffusion models to support additional input conditions. The ControlNet learns task-specific conditions in an end-to-end way, and the learning is robust even when the training dataset is small (< 50k). Moreover, training a ControlNet is as fast as fine-tuning a diffusion model, and the model can be trained on a personal devices. Alternatively, if powerful computation clusters are available, the model can scale to large amounts (millions to billions) of data. We report that large diffusion models like Stable Diffusion can be augmented with ControlNets to enable conditional inputs like edge maps, segmentation maps, keypoints, etc. This may enrich the methods to control large diffusion models and further facilitate related applications. Adding Conditional Control to Text-to-Image Diffusion Models | Lvmin Zhang, Maneesh Agrawala

There exists a very subtle but powerful methodology for deep learning. This simple idea is to use a neural network to extract the “specification” from an example. Then use another neural network to generate examples from the extracted example. This is a universal idea. Deep Learning Modularity and Language Models | Carlos E. Perez - Medium

he.png