MATH 103: Multivariate Calculus for Machine Learning
| Course Code | MATH 103 |
| Course Name | Multivariate Calculus for Machine Learning |
| Department | Mathematics |
| Semester Offered | Odd (Term 1) |
| Tuition Hours | 21 hours |
| Course Level | Foundational |
| Pre-requisite | MATH 102: Calculus for Machine Learning |
| Co-requisite | MATH 101: Linear Algebra for Machine Learning |
| Course Objective | Single-variable calculus tells you how one thing changes. That is not enough. Real systems do not depend on one variable. They depend on hundreds or thousands of parameters interacting together. This course extends calculus into that reality. Students will learn how change works in high-dimensional systems, how gradients guide optimization across many parameters, and how modern machine learning models actually navigate complex loss landscapes. By the end, students should be able to look at a training process and understand how multiple parameters are being updated together, not just mechanically but conceptually. This is where calculus stops being academic and starts becoming the engine behind real learning systems. |
| Course Philosophy | This course emphasizes
|
| Course Learning Outcomes | Upon successful completion of this course, students will be able to:
|
| Course Author | Sagar Udasi MSc Statistics and Data Science with Computational Finance from The University of Edinburgh. Contact: sagar.l.udasi@gmail.com |
| Course Organiser | TBD |
| No. | Lecture Title | Concepts Covered | Lecture Objective |
|---|---|---|---|
| 01 | When One Variable Is Not Enough | Multivariable functions, surfaces | Students understand why real ML models depend on many variables, not just one. |
| 02 | Visualizing Loss in 3D and Beyond | Surface plots, intuition of high-dimensional spaces | Helps students build intuition for optimization landscapes used in training. |
| 03 | Partial Derivatives: Changing One Thing at a Time | Partial derivatives, interpretation | Students learn how individual parameters affect output in a system. |
| 04 | Gradients: The Direction That Actually Matters | Gradient vector, geometric meaning | Core concept for optimization in ML. Students understand direction of steepest change. |
| 05 | Gradient Descent in High Dimensions | Multi-variable optimization | Extends earlier knowledge to real ML systems with many parameters. |
| 06 | When Optimization Gets Tricky | Saddle points, flat regions | Prepares students to deal with real-world training difficulties. |
| 07 | The Chain Rule Scales Up | Multivariable chain rule | Foundation for understanding backpropagation in deep networks. |
| 08 | What Backpropagation Really Does in High Dimensions | Gradient flow through layers | Connects multivariate calculus directly to neural network training. |
| 09 | Jacobians Without the Fear | Jacobian matrix, transformations | Helps students understand how outputs change with respect to multiple inputs. |
| 10 | Constraints Change Everything | Constrained optimization, Lagrange multipliers (intuition) | Introduces real-world limitations in optimization problems. |
| 11 | Regularization: Controlling the Model | Penalties, overfitting intuition | Connects calculus to better generalization in ML systems. |
| 12 | From Math to Code | Implementing multivariate gradient descent | Students translate theory into working optimization routines. |
| 13 | Lab: Training a Multi-Parameter Model | Hands-on optimization | Students apply concepts to a real training loop. |
| 14 | Project Integration: Tuning Your AI System | Applying optimization to product | Students directly improve their Term 1 AI agent using learned concepts. |
| Component | Weightage |
|---|---|
| Written Examination (2 hours) | 50% |
| Practical Assignment (Optimization Task) | 30% |
| Project Integration (Applied to AI Product) | 20% |
| Type | Resource | Provider |
|---|---|---|
| Lecture | Multivariable Calculus | MIT OpenCourseWare |
| Lecture | Essence of Calculus (Multivariable Sections) | 3Blue1Brown |
| Reading | Calculus: Early Transcendentals (Multivariable Chapters) | James Stewart |
| Practical | CS231n Optimization and Backprop Notes | Stanford University |