生成AIの技術詳細(181~210)– 生成モデルの内部構造と技術を深く理解します。 –
-
Chapter 7
[AI from Scratch] Episode 232: Practical Transfer Learning — Utilizing Pre-trained Models
Recap and Today's Theme Hello! In the previous episode, we explored data augmentation, demonstrating how to expand image datasets using Keras. We confirmed that data augmentation not only helps models learn from more diverse data but als... -
Chapter 7
[AI from Scratch] Episode 231: Practical Data Augmentation — Expanding Image Datasets
Recap and Today's Theme Hello! In the previous episode, we covered saving and loading models using Keras, learning how to save trained models for future use. This helps streamline both development and deployment processes. Today, we will... -
Chapter 7
[AI from Scratch] Episode 209: Neural Radiance Fields (NeRF)
Recap: Diffusion Models In the previous episode, we discussed Diffusion Models, explaining their mechanisms, applications, and differences from other generative models like GANs and VAEs. Diffusion models are gaining attention for their ... -
Chapter 7
[AI from Scratch] Episode 208: Diffusion Models — An Introduction to Diffusion Models
Recap: Challenges and Limitations of Generative Models In the previous episode, we explored the challenges and limitations of generative models, including quality issues, computational costs, and ethical concerns. While generative models... -
Chapter 7
[AI from Scratch] Episode 207: Challenges and Limitations of Generative Models
Recap: Applications of Generative Models In the previous episode, we explored various applications of generative models such as image generation, text generation, and speech synthesis. These technologies have broad applications, ranging ... -
Chapter 7
[AI from Scratch] Episode 206: Applications of Generative Models — Image Generation, Text Generation, and Speech Synthesis
Recap: Model Safety and Filtering In the previous episode, we explored filtering techniques to ensure the safety of AI models. Methods such as the blacklist approach, NLP filtering, and Human-in-the-Loop (HITL) were discussed to prevent ... -
Chapter 7
[AI from Scratch] Episode 205: Model Safety and Filtering — Preventing Inappropriate Outputs
Recap: Prompt Tuning In the previous episode, we learned about prompt tuning, a technique to optimize prompts for eliciting desired outputs from pre-trained models. Designing prompts effectively is crucial for enhancing model accuracy an... -
Chapter 7
[AI from Scratch] Episode 204: Prompt Tuning — Optimizing Prompts to Enhance Model Performance
Recap: Large-Scale Pre-Trained Models In the previous episode, we discussed large-scale pre-trained models, which are models trained on massive datasets that exhibit high performance when their general features are applied to specific ta... -
Chapter 7
[AI from Scratch] Episode 203: Large-Scale Pre-Trained Models — Advantages and Applications of Pre-Trained Models
Recap: Applications of Self-Supervised Learning In the previous episode, we discussed the applications of self-supervised learning, a method that learns features from unlabeled data. This approach has proven valuable in fields such as na... -
Chapter 7
[AI from Scratch] Episode 202: Applications of Self-Supervised Learning — Learning from Unlabeled Data
Recap: Evaluation Metrics for Speech Generation In the previous episode, we explained how to evaluate the quality of speech generation using metrics like PESQ and STOI for objective evaluation and MOS for subjective evaluation. These met... -
Chapter 7
[AI from Scratch] Episode 201: Evaluation Metrics for Speech Generation — PESQ, STOI, and More
Recap: Tacotron In the previous episode, we explained Tacotron, a model that converts text into speech, widely used in applications such as voice assistants and narration generation. Especially with Tacotron 2, the quality of the generat... -
Chapter 7
[AI from Scratch] Episode 200: Tacotron — A Model for Text-to-Speech Conversion
Recap: WaveNet In the previous episode, we explained WaveNet, a neural network-based model that directly generates speech waveforms, offering high-quality speech synthesis. By generating audio at the sample level, WaveNet produces more n... -
Chapter 7
[AI from Scratch] Episode 198: Speech Generation Models — Basics of Speech Synthesis Technology
Recap: Evaluation Metrics for Text Generation In the previous episode, we discussed evaluation metrics for text generation, explaining key metrics like Perplexity and the BLEU score. Perplexity measures a model's prediction accuracy, whi... -
Chapter 7
[AI from Scratch] Episode 199: WaveNet — Explaining the High-Quality Speech Generation Model
Recap: Speech Generation Models In the previous episode, we discussed the basics of speech generation technology. Traditional methods such as rule-based, unit selection, and parametric speech synthesis have evolved into modern approaches... -
Chapter 7
[AI from Scratch] Episode 197: Evaluation Metrics for Text Generation — Perplexity and BLEU Score
Recap: BERT and the Masked Language Model In the previous episode, we explored BERT (Bidirectional Encoder Representations from Transformers), a powerful model in natural language processing (NLP), and its training method, the Masked Lan...
