0% found this document useful (0 votes)
7 views4 pages

Deep Learning - Complete Notes

Uploaded by

prachiksk07
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views4 pages

Deep Learning - Complete Notes

Uploaded by

prachiksk07
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
You are on page 1/ 4

Here’s a **complete and in-depth note on Deep Learning (DL)** — written in clear,

structured points for easy understanding and exam or interview use 👇

---

# 🧠 DEEP LEARNING – COMPLETE NOTES

---

1## 1 What is Deep Learning?


1️⃣

**Definition:**
Deep Learning (DL) is a **subset of Machine Learning (ML)** that uses **Artificial
Neural Networks (ANNs)** with **multiple hidden layers** to automatically learn
complex patterns from large amounts of data.

**Key idea:**
It tries to **mimic how the human brain works** — by processing data through layers
of interconnected “neurons.”

---

2️⃣
## Why “Deep”?

The term **“deep”** refers to the **number of layers** in a neural network.

* Traditional ML: Few layers (shallow learning).


* Deep Learning: Many hidden layers that allow automatic feature extraction and
abstraction.

Example:

```
Input → Hidden Layer 1 → Hidden Layer 2 → Hidden Layer 3 → Output
```

---

3️⃣
## Difference between Machine Learning and Deep Learning

| Feature | Machine Learning | Deep Learning |


| ------------------ | --------------------- | ------------------------------ |
| Data requirement | Works with small data | Requires huge data |
| Feature extraction | Manual (by experts) | Automatic (learned by network) |
| Execution time | Faster to train | Slower (high computation) |
| Hardware | Runs on CPU | Needs GPU/TPU |
| Example | Decision Tree, SVM | CNN, RNN, Transformer |

---

4️⃣
## Biological Inspiration

Deep Learning is inspired by the **human brain’s neurons**:

* Each **neuron** receives inputs, applies weights, adds bias, and passes through
an **activation function**.
* The result is transmitted to the next layer, simulating how the brain processes
signals.
---

5️⃣
## Structure of a Neural Network

### Basic Elements:

1. **Input Layer** – Takes the features (e.g., pixel values of an image).


2. **Hidden Layers** – Perform transformations and extract patterns.
3. **Output Layer** – Produces the final prediction/class.

### Example:

For image recognition of a cat:

```
Input: Image pixels
Hidden layers: Feature extraction (edges, shapes, textures)
Output: “Cat”
```

---

6️⃣
## Key Concepts

### 1. **Neuron / Perceptron**

* Mathematical model of a biological neuron.


* Formula:
[
y = f(w_1x_1 + w_2x_2 + ... + w_nx_n + b)
]
where
(w) = weights, (x) = inputs, (b) = bias, (f) = activation function.

---

### 2. **Activation Function**

Adds **non-linearity** to the model so it can learn complex patterns.

| Function | Formula | Characteristics


|
| ---------- | ------------------------- |
------------------------------------------ |
| Sigmoid | 1 / (1 + e^-x) | Output between 0–1, good for
probabilities |
| Tanh | (e^x - e^-x)/(e^x + e^-x) | Output between -1 to 1
|
| ReLU | max(0, x) | Fast and popular for hidden layers
|
| Leaky ReLU | x if x>0 else 0.01x | Fixes ReLU “dead neuron” issue
|
| Softmax | e^xᵢ / Σ e^xⱼ | Used in output layer for classification
|

---

### 3. **Loss Function**


Measures **how far predictions are from actual output**.

| Task | Example Loss Function |


| -------------- | ------------------------ |
| Regression | Mean Squared Error (MSE) |
| Classification | Cross Entropy / Log Loss |

---

### 4. **Optimization**

Used to minimize loss function.

* **Gradient Descent**: Main technique for optimization.


Updates weights in opposite direction of the gradient.

[
w_{new} = w_{old} - \eta \times \frac{dL}{dw}
]

where ( \eta ) = learning rate.

* Variants:

* Stochastic Gradient Descent (SGD)


* Mini-batch Gradient Descent
* Adam Optimizer (adaptive)

---

### 5. **Backpropagation**

* Algorithm to **update weights** based on error propagation.


* Steps:

1. Forward pass → Compute output and loss.


2. Backward pass → Compute gradients (errors).
3. Update weights → Using optimizer.

---

7️⃣
## Types of Deep Learning Networks

### 1. **Feedforward Neural Network (FNN)**

* Simplest type, where data moves in one direction: Input → Output.


* No feedback loops.

### 2. **Convolutional Neural Network (CNN)**

* Used for **image/video** data.


* Automatically extracts features like edges, textures, shapes.
* Components:

* Convolution Layer
* Pooling Layer
* Fully Connected Layer
* Examples: VGGNet, ResNet, AlexNet, LeNet.
### 3. **Recurrent Neural Network (RNN)**

* Used for **sequential data** (text, time series).


* Maintains memory of previous inputs using feedback loops.
* Types:

* Vanilla RNN
* LSTM (Long Short-Term Memory)
* GRU (Gated Recurrent Unit)
* Example: Sentiment analysis, speech recognition.

### 4. **Autoencoders**

* Used for **data compression or noise reduction**.


* Structure:

* Encoder: Compress data


* Decoder: Reconstruct original
* Example: Denoising images.

### 5. **Generative Adversarial Networks (GANs)**

* Two networks:

* **Generator** – creates fake data.


* **Discriminator** – detects if data is real or fake.
* Example: Deepfake videos, art generation.

### 6. **Transformers**

* Replaces RNNs for NLP.


* Uses **attention mechanism** to focus on relevant parts of input.
* Examples: BERT, GPT (ChatGPT uses Transformer architecture).

---

8️⃣
## Training Process in Deep Learning

1. **Collect Data**
2. **Preprocess Data** (normalization, resizing, tokenization)
3. **Define Model Architecture**
4. **Choose Loss Function & Optimizer**
5. **Train Model** (Forward + Backward propagation)
6. **Validate & Tune Hyperparameters**
7. **Test & Deploy**

---

9️⃣
## Hyperparameters in Deep Learning

| Hyperparameter | Meaning |
| -------------- | -------------------------------- |
| Learning Rate | Controls how fast weights update |
| Batch Size | Number of samples per update |
| Epochs | Number of passes through dataset |
| Number | |

You might also like