How Forward Propagation Works:
z = (x_1 w_1 + x_2 w_2 + ... + x_n w_n) + bImportance:
How Backpropagation Works:
| Aspect | Forward Propagation | Backpropagation |
|---|---|---|
| Direction of Flow | Input → Hidden layers → Output layer | Output → Hidden layers → Input layer |
| Main Purpose | Generate predictions from inputs | Learn by correcting errors |
| When it Happens | First step in training and testing | After forward propagation during training |
| Uses Target Output? | No | Yes |
| Error Calculation | Does not calculate error | Uses loss function to compute error |
| Weight Update | Weights remain unchanged | Weights are updated |
| Mathematical Role | Computes activations using weights and bias | Computes gradients using chain rule |
| Learning Involved | No learning happens | Learning happens here |
| Optimization Needed | No | Yes (Gradient Descent, Adam, etc.) |
| Example Meaning | “What is the output?” | “How can we improve the output?” |
(5^2 = 25)Optimization techniques are methods used in machine learning and deep learning to improve a model’s performance.
Their main goal is to reduce the loss (error) and make the model’s predictions more accurate.
During training, the model repeatedly updates its weights using an optimizer until it reaches the best possible solution, called the minimum loss.
Stochastic Gradient Descent (SGD):
Adam Optimizer:
∣ 𝑤 ∣𝑤²0.2–0.5Made By SOU Student for SOU Students