Thanks to improvements in computing power and scientific theory, Generative AI is more accessible than ever before. Generative AI will play a significant role across industries and will gain significant importance due to its numerous applications such as Creative Content Generation, Data Augmentation, Simulation and Planning, Anomaly Detection, Drug Discovery, and Personalized Recommendations etc. In this course we will take a deeper dive on denoising diffusion models, which are a popular choice for text-to-image pipelines, disrupting several industries.

 

Learning Objectives


By participating in this workshop, you’ll learn how to:
  • Build a U-Net to generate images from pure noise
  • Improve the quality of generated images with the Denoising Diffusion process
    • Compare Denoising Diffusion Probabilistic Models (DDPMs) with Denoising Diffusion Implicit Models (DDIMs)
  • Control the image output with context embeddings
  • Generate images from English text-prompts using CLIP

Datasheet Coming Soon

Workshop Outline

Introduction
(15 mins)
From U-Nets to Diffusion
(60 mins)
  • Build a U-Net, a type of autoencoder for images.
  • Learn about transposed convolution to increase the size of an image.
  • Learn about non-sequential neural networks and residual connections.
  • Experiment with feeding noise through the U-Net to generate new images
Break (10 mins)
Control with Context
(60 minutes)
  • Learn how to alter the output of the diffusion process by adding context embeddings
  • Add additional model optimizations such as
  • Sinusoidal Position Embeddings
  • The GELU activation function
  • Attention
Text-to-Image with CLIP
(60 minutes)
  • Walk through the CLIP architecture to learn how it associates image embeddings with text embeddings
  • Use CLIP to train a text-to-image diffusion model
Break (60 mins)
State-of-the-art Models
(60 mins)
  • Review various state-of-the-art generative ai models and connect them to the concepts learned in class
  • Discuss prompt engineering and how to better influence the output of generative AI models
  • Learn about content authenticity and how to build trustworthy models
Final Review
(60 mins)
  • Review key learnings and answer questions.
  • Complete the assessment and earn a certificate.
  • Complete the workshop survey.
  • Learn how to set up your own AI application development environment.
Next Steps
Continue learning with these DLI trainings:
 

Workshop Details

Duration: 8 hours

Price: $500 for public workshops, contact us for enterprise workshops.

Prerequisites:

Technologies: PyTorch, CLIP

Hardware: Desktop or laptop computer capable of running the latest version of Chrome or Firefox. Each participant will be provided with dedicated access to a fully configured, GPU-accelerated workstation in the cloud.

Certificate: Upon successful completion of the assessment, participants will receive an NVIDIA DLI certificate to recognize their subject matter competency and support professional career growth.

Languages: English

Upcoming Public Workshops

If your organization is interested in boosting and developing key skills in AI, accelerated data science, or accelerated computing, you can request instructor-led training from the NVIDIA DLI.

Continue Your Learning with These DLI Trainings

Getting Started with Image Segmentation

Modeling Time-Series Data with Recurrent Neural Networks in Keras

Building Transformer-Based Natural Language Processing Applications

Fundamentals of Deep Learning for Multi-GPUs