Deep Learning Roadmap (Book +
Practice Integrated)
Stage 1: Core Neural Networks
- Biological inspiration & history.
- Perceptrons, MLPs, forward/backprop, loss functions.
- Activations (Sigmoid, Tanh, ReLU, Softmax).
- Gradient descent variants (SGD, Mini-batch, Momentum, Adam).
- Project: Implement NN from scratch (NumPy).
Stage 2: Regularization & Optimization
- L1, L2, dropout, early stopping.
- Weight initialization (Xavier, He).
- Batch Norm, Layer Norm.
- Dealing with vanishing/exploding gradients.
- Practical debugging & hyperparameter tuning.
- Project: MNIST classifier with/without regularization.
Stage 3: CNNs & Computer Vision
- Convolutions, filters, pooling, CNN stack.
- Classic architectures: LeNet, AlexNet, VGG, Inception, ResNet.
- Data augmentation, transfer learning.
- Advanced: object detection (YOLO, Faster R-CNN), segmentation (U-Net).
- Project: CIFAR-10 classifier, transfer learning on dogs vs. cats.
Stage 4: Sequence Modeling & NLP
- RNNs, LSTMs, GRUs.
- Seq2Seq encoder-decoder.
- Attention mechanisms.
- Applications: sentiment analysis, machine translation.
- Project: LSTM text generator, Seq2Seq translator.
Stage 5: Transformers & Large Models
- Transformer architecture: self-attention, multi-head, positional encodings.
- BERT, GPT, T5, RoBERTa.
- Pretraining vs fine-tuning, distillation & quantization.
- Efficient transformers (Reformer, Linformer).
- Projects: fine-tune BERT for Q&A, summarizer using T5.
Stage 6: Representation Learning & Research
- Autoencoders: denoising, variational, sparse.
- GANs & VAEs.
- Structured probabilistic models.
- Monte Carlo methods, Approximate inference.
- Energy-based models, Boltzmann machines.
- Advanced generative models.
- Projects: Variational Autoencoder on MNIST, GAN for image generation.