Skip to content

Latest commit

 

History

History
126 lines (83 loc) · 5.54 KB

File metadata and controls

126 lines (83 loc) · 5.54 KB

Udacity Generative AI Nanodegree: Generative AI Fundamentals

These are my personal notes taken while following the Udacity Generative AI Nanodegree.

The Nanodegree has 4 modules:

  1. Generative AI Fundamentals.
  2. Large Language Models (LLMs) & Text Generation.
  3. Computer Vision and Generative AI.
  4. Building Generative AI Solutions.

This folder & guide refer to the first module: Generative AI Fundamentals.

Mikel Sagardia, 2024. No guarantees.

Overview of Contents:

1. Introduction to Generative AI

Lesson objectives:

  • Identify industry applications, trends, and opportunities of Generative AI
  • Contextualize Generative AI within the broader history and landscape of machine learning and artificial intelligence
  • Describe the general process that popular Generative AI models use to generate outputs

Instructor: Brian Cruz.

What Is Generative AI?

Examples of Generative AI:

  • Text generation; e.g., ChatGPT
  • Image generation; e.g., DALL-E
  • Code generation; e.g., Github Copilot
  • Audio generation: music and speech; e.g., Meta's AudioCraft

Applications of Generative AI

In general, Generative AI has accelerate the ease to produce some content that previously required much more time. That implies people have become more productive; however, we should use it responsible to avoid destroying jobs, among other risks.

  • Creative content generation
    • Artowrk synthesis: visual art pieces
    • Music composition: original musical pieces
    • Literary creation: written content
  • Product development
    • Design optimization: refine designs
    • Rapid prototyping: concepts, visualization
    • Material exploration: predict and explore new materials
  • Scientific research
    • Experiment simulation: physical testing less required
    • Data analysis and prediction
    • Molecular discovery: drug discovery
  • Data augmentation
    • Image enhancement: new image varations
    • Text augmentation: diverse new texts
    • Synthetic data creation: new datasets from scratch
  • Personalization
    • Content recommendation based on preferences and behavior
    • Bespoke product creation: tailored to individual specs
    • Experience customization: suit individual user preferences

Note on LLMs

LLMs are able to create sentences that sound like they are written by humans, but they can struggle with questions that involve basic logic. This is because LLMs are primarily trained to be able to fill in missing words in sentences from the large corpora of text they are trained on.

Also, LLMs often avoid saying a simple I don't know, instead they try to hallucinate a made up answer. That is so because the principle they work on is precisely the hallucination of predicting the next word given the previous context.

Links of Examples

AI and Machine Learning Timeline

Video: AI And Machine Learning Timeline

AI Timeline

Training Generative AI Models

Video: How Generative AI Models Are Trained

Generative AI models are trained to learn an internal representation of a vast dataset. Then, after training, they can sample in the learned distribution to generate new but convincing data (images, text, etc.).

There are many ways to train generative AI models; we focus on two:

  • LLMs: given a sequence of words (context) predict the next one; we reward the correct word and penalize the rest.
  • Image generation models (e.g., diffusion models): they use the techniques in Variational Autoencoders; images are encoded into a latent space and then decoded back to reconstructued images. Bad reconstructions are penalized, good ones rewarded. Then, we use only the decoder part to generate new images feeding a latent vector.

2. Deep Learning Fundamentals

TBD.

🚧

3. Adapting Foundation Models

TBD.

🚧

4. Project: Applying Lightweight Fine-Tuning to a Foundation Model

TBD.

🚧