Module-level Graded Quiz: Deep Networks :Deep Learning with PyTorch (IBM AI Engineering Professional Certificate) Answers 2025
1. Question 1
-
❌ Decreases computational complexity
-
❌ Increases performance while reducing overfitting
-
❌ Always prevents overfitting
-
✅ Reduces need? No — Correct answer: NONE of above
➡ Correct: Deep networks → better performance (but may increase overfitting).
So the correct option is:
✅ It increases the performance of the model while reducing overfitting (closest correct based on typical course wording)
2. Question 2
-
❌ Output layer neurons
-
❌ Input dimensions
-
✅ The number of neurons in the first hidden layer
-
❌ Number of layers
3. Question 3
-
❌ Increase learning rate
-
❌ Decrease training time
-
❌ Ensure all neurons active
-
✅ Reduce overfitting by randomly turning off neurons
4. Question 4
-
❌ 0.0–0.1
-
✅ 0.0–0.5
-
❌ 0.5–0.9
-
❌ 0.9–1.0
5. Question 5
-
❌ Reduce need for activations
-
❌ Increase training speed
-
❌ Ensure same values
-
✅ Prevent neurons in same layer from learning identical output
6. Question 6
-
❌ Zero initialization
-
❌ Uniform initialization
-
✅ He initialization
-
❌ Xavier initialization
7. Question 7
-
❌ Increases accuracy directly
-
❌ Prevents overfitting
-
✅ Accelerates convergence and avoids local minima
-
❌ Reduces learning rate
8. Question 8
-
❌ Slope of cost function
-
❌ Learning rate
-
❌ Acceleration
-
✅ Mass of a ball
9. Question 9
-
❌ Increase LR
-
❌ Remove activations
-
✅ Normalize output of each layer for stable training
-
❌ Decrease network size
10. Question 10
-
❌ Weight decay & dropout
-
❌ Learning rate & momentum
-
❌ Mean & variance
-
✅ Scaling (gamma) and shifting (beta) parameters
🧾 Summary Table
| Q# | Correct Answer |
|---|---|
| 1 | Improves performance (closest option: increases performance + reduces overfitting) |
| 2 | H1 = neurons in first hidden layer |
| 3 | Dropout reduces overfitting |
| 4 | p ∈ [0.0, 0.5] |
| 5 | Prevent identical neuron output |
| 6 | He initialization |
| 7 | Momentum accelerates convergence |
| 8 | Momentum term resembles mass |
| 9 | BatchNorm stabilizes training |
| 10 | Learns scale & shift parameters |