๐”– Scriptorium
โœฆ   LIBER   โœฆ

๐Ÿ“

Learn Python Generative AI: Journey from autoencoders to transformers to large language models

โœ Scribed by Zonunfeli Ralte, Indrajit Kar


Publisher
BPB Publications
Year
2024
Tongue
English
Leaves
348
Category
Library

โฌ‡  Acquire This Volume

No coin nor oath required. For personal study only.

โœฆ Synopsis


Learn to unleash the power of AI creativity KEY FEATURES โ— Understand the core concepts related to generative AI. โ— Different types of generative models and their applications. โ— Learn how to design generative AI neural networks using Python and TensorFlow. DESCRIPTION This book researches the intricate world of generative Artificial Intelligence, offering readers an extensive understanding of various components and applications in this field. The book begins with an in-depth analysis of generative models, providing a solid foundation and exploring their combination nuances. It then focuses on enhancing TransVAE, a variational autoencoder, and introduces the Swin Transformer in generative AI. The inclusion of cutting edge applications like building an image search using Pinecone and a vector database further enriches its content. The narrative shifts to practical applications, showcasing GenAI's impact in healthcare, retail, and finance, with real-world examples and innovative solutions. In the healthcare sector, it emphasizes AI's transformative role in diagnostics and patient care. In retail and finance, it illustrates how AI revolutionizes customer engagement and decision making. The book concludes by synthesizing key learnings, offering insights into the future of generative AI, and making it a comprehensive guide for diverse industries. Readers will find themselves equipped with a profound understanding of generative AI, its current applications, and its boundless potential for future innovations. WHAT YOU WILL LEARN โ— Acquire practical skills in designing and implementing various generative AI models. โ— Gain expertise in vector databases and image embeddings, crucial for image search and data retrieval. โ— Navigate challenges in healthcare, retail, and finance using sector specific insights. โ— Generate images and text with VAEs, GANs, LLMs, and vector databases. โ— Focus on both traditional and cutting edge techniques in generative AI. WHO THIS BOOK IS FOR This book is for current and aspiring emerging AI deep learning professionals, architects, students, and anyone who is starting and learning a rewarding career in generative AI

โœฆ Table of Contents


Table of Contents

  1. Introducing Generative AI

Introduction

Structure

Objectives

Overview of generative models

Discriminative vs. generative models

Types of discriminative and generative models

Strengths and weaknesses

Class imbalance scenario

Generative modeling framework

Sample Space

Probability density function

Maximum likelihood

KL divergence

GMM code using TensorFlow probability

Conclusion

  1. Designing Generative Adversarial Networks

Introduction

Structure

Objectives

Generative Adversarial Networks

Types of GANs available

Architecture of a GAN

Equation

Discriminator loss

Generator loss

Vanilla GAN

Outline crucial factors in GAN architecture design

Major challenges in designing GANs architecture

Architecture of Deep Convolutional GANs

Architecture of Wasserstein GANs

Architecture of Conditional GANs

Architecture of CycleGANs

Architecture of progressive GANs

Architecture of StyleGANs

Architecture of Pix2Pix

Conclusion

Multiple choice questions

Answers

  1. Training and Developing Generative Adversarial Networks

Introduction

Structure

Objectives

Generative Adversarial Training

Generating MNIST data: Basic GAN implementation

Issues during training a GANs

Mode collapse

Vanishing gradients

Oscillation

Unstability

Evaluation

Case study: Common practical implementation of GANs for augmentation and balancing classes

Conclusion

  1. Architecting Auto Encoder for Generative AI

Introduction

Structure

Objectives

Auto Encoders

Regularization

Creating a bottleneck

Key distinctions with autoencoders

Autoencoders

GANs

Importance of regularization in auto encoders

Cifar10

Anomaly detection using auto encoder

Autoencoders with convolutional layers

Architecture

Capturing spatial information

CNN versus ANN Autoencoders

Conclusion

  1. Building and Training Generative Autoencoders

Introduction

Structure

Objectives

Latent space

Difference between GANs latent space and AE latent space

Key distinctions with autoencoders latent space

Adding color to a grayscale image using autoencoders

Coding advanced auto encoders

Multi modal auto encoders

Loss in autoencoders

Mean squared error loss

Binary cross-entropy loss

Categorical cross-entropy loss

Kullback-leibler divergence loss

Huber loss

Challenges in training auto encoders and mitigation

AE vs. VAE

Conclusion

  1. Designing Generative Variation Auto Encoder

Introduction

Structure

Objectives

Story of VAE

VAE vs AE

Math behind the latent space

Deterministic Autoencoder

Stochastic Variational Autoencoder

Key distinctions with autoencoder latent space

Can the VAE Latent space be stochastic as well as deterministic

Dirichlet distribution

Importance of the latent space when designing a VAE

Vanilla VAE architecture

The ELBO

The reparameterization trick

Challenges in Vanilla VAE

Types of VAE

Conclusion

  1. Building Variational Autoencoders for Generative AI

Introduction

Structure

Objectives

Key focus areas in VAE research

Building a VAE with Dirichlet distribution: Non-CNN Approach

Building a VAE with Dirichlet distribution: CNN Approach

Difference between two networks

VAE with Non Dirichlet distribution

KL divergence

Common loss function sin VAE

Common issues and possible solutions while training VAE

Missing data handling during generation

Optimization techniques

Conclusion

  1. Fundamental of Designing New Age Generative Vision Transformer

Introduction

Structure

Objectives

The evolution

The birth of transformers

Overview of transformer architectures

Applications in NLP

Generative transformers and language modeling

Transformer in computer vision

Difference between VAE, GANs, and Transformers

Transformers

Generative Adversarial Networks

Variational autoencoders

Differences and applications

Vision Transformer

Understanding self-attention

NLP vs vision

NLP transformer

Self-attention mechanism

Feed-forward neural networks

Vision transformer

Patch embeddings

Positional embeddings

Transformer encoder

Architectural attention

Dot product attention

Scaled dot product attention

Additive attention

Multi-head attention

Cross attention

Compute attention scores

Compute cross-attention output

When to use which architectural attention

Functional attention

Hard attention

Equation: Sampling-based hard attention

Soft attention

Equation: Soft attention

Global attention

Equation: Global attention

Local attention

Equation: local attention

When to use which functional attention

Hard attention

Soft attention

Global attention

Local attention

Conclusion

  1. Implementing Generative Vision Transformer

Introduction

Structure

Objectives

STL dataset

Key features of the STL-10 dataset

Developing a VAE model on STL dataset

Implementation of VAE architecture with TensorFlow

Outputs

Pytorch

Transition from VAE to Generative Transformer Model: Keras Vit Library

Implementing a ViT model from scratch

Outputs

Implementing a ViT model pre trained with ViT model

Outputs

Training Pretrained ViT vs ViT scratch

Pretrained Vision Transformer

Advantages

Disadvantages

Training a VIT model from scratch

Advantages

Disadvantages

Examining the loss curve

Optimization of ViT models

Conclusion

  1. Architectural Refactoring for Generative Modeling

Introduction

Structure

Objectives

STL dataset

Exploring the combination process: Outline

Refactoring TransVAE and improving

Cyclic Learning Rate Schedule

LearningRateScheduler

EarlyStopping

Weight decay: L2 regularization

Improved Encoder Decoder

SWIN-Transformer

Implementation of SWIN Transformer: VAE

Improving the models

Conclusion

  1. Major Technical Roadblocks in Generative AI and Way Forward

Introduction

Structure

Objectives

Challenges and hurdles in Generative AI

NLP based generative models

Large language models and image-based foundation models

Embedding in language models

Embedding in image

Generative AI and embeddings

Vector data bases and image embeddings

Vector databases

Image embeddings

Building an image search using pinecone and vector database

Conclusion

  1. Overview and Application of Generative AI Models

Introduction

Structure

Objectives

GenAI in hospital

GenAI in dental

GenAI in radiology

GenAI in retail

GenAI in finance

GenAI in corporate finance

GenAI in insurance

Conclusion

  1. Key Learnings

Introduction

Structure

Objectives

Key learning from all the chapters

Chapter 1: Introducing Generative AI

Chapter 2: Designing Generative Adversarial Networks

Chapter 3: Training and Developing Generative Adversarial Networks

Chapter 4: Architecting Auto Encoder for Generative AI

Chapter 5: Building and Training Generative Autoencoders

Chapter 6: Designing Generative VAE

Chapter 7: Building Variational AutoEncoders for Generative AI

Chapter 8: Designing New Age Generative Vision Transformer for Generative Learning

Chapter 9: Implementing Generative Vision Transformers

Chapter 10: Architectural Refactoring Combining Encoder-decoder and Transformers for Generative Modeling

Chapter 11: Major Technical Roadblocks in Generative AI

Chapter 12: Overview of Applications of Generative AI Models

Index


๐Ÿ“œ SIMILAR VOLUMES


Learn Python Generative AI: Journey from
โœ Zonunfeli Ralte, Indrajit Kar ๐Ÿ“‚ Library ๐Ÿ“… 2024 ๐Ÿ› BPB Publications ๐ŸŒ English

Learn to unleash the power of AI creativity KEY FEATURES โ— Understand the core concepts related to generative AI. โ— Different types of generative models and their applications. โ— Learn how to design generative AI neural networks using Python and TensorFlow. DESCRIPTION This book researches the intri

LLM, Transformer, RAG AI: Mastering Larg
โœ Code, Et Tu ๐Ÿ“‚ Library ๐Ÿ“… 2024 ๐ŸŒ English

Explore the world of language models with "LLM, Transformer, RAG AI: Mastering Large Language Models, Transformer Models, and Retrieval-Augmented Generation (RAG) Technology." Dive into the fundamentals of language model development, from Natural Language Processing basics to choosing the right fram

Large Language Model-Based Solutions : H
โœ Shreyas Subramanian ๐Ÿ“‚ Library ๐Ÿ“… 2024 ๐Ÿ› WILEY ๐ŸŒ English

<p><b>Learn to build cost-effective apps using Large Language Models</b> <p>In <i>Large Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications</i>, Principal Data Scientist at Amazon Web Services, Shreyas Subramanian, delivers a practical guide for develo

Large Language Model-Based Solutions: Ho
โœ Shreyas Subramanian ๐Ÿ“‚ Library ๐Ÿ“… 2024 ๐Ÿ› WILEY ๐ŸŒ English

Learn to build cost-effective apps using Large Language Models InLarge Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications, Principal Data Scientist at Amazon Web Services, Shreyas Subramanian, delivers a practical guide for developers and data scient

Generative AI with LangChain: Build larg
โœ Ben Auffarth ๐Ÿ“‚ Library ๐Ÿ“… 2023 ๐Ÿ› Packt ๐ŸŒ English

ChatGPT and the GPT models by OpenAI have brought about a revolution not only in how we write and research but also in how we can process information. This book discusses the functioning, capabilities, and limitations of LLMs underlying chat systems, including ChatGPT and Bard. It also demonstrates,

Large Language Model-Based Solutions: Ho
โœ Shreyas Subramanian ๐Ÿ“‚ Library ๐Ÿ“… 2024 ๐Ÿ› Wiley ๐ŸŒ English

<p><span>Learn to build cost-effective apps using Large Language Models</span></p><p><span>In </span><span>Large Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications</span><span>, Principal Data Scientist at Amazon Web Services, Shreyas Subramanian, del