Lectures¶
Complete lecture materials with slides, review questions, theory connections, and PhD-level insights — all in PDF format.
Course Administration¶
Assessment Breakdown¶
| Component | Weight | Details |
|---|---|---|
| Quizzes | 40% | 4-5 quizzes throughout the course |
| Midterm Examination | 20% | Covers Part I: Mathematical Foundations |
| Final Examination | 30% | Comprehensive exam covering all topics |
| Reading & Review Papers | 10% | 4-5 research papers to review |
| TOTAL | 100% | - |
Course Level & Prerequisites¶
Target Audience: Undergraduate/Graduate Level
Flexible Prerequisites:
- Basic linear algebra and vector calculus recommended
- Course structure allows for adjustment based on student background
- Mathematical concepts presented with varying levels of rigor
Lecture Structure¶
Every lecture follows a consistent four-part approach designed to build deep understanding:
| Part | What It Covers |
|---|---|
| Concepts & Explanation | Key definitions, intuitive explanations, "Think of it as..." plain-English descriptions |
| Mathematical Examples & Tutorials | Step-by-step worked examples with detailed solutions |
| Review & Practice | 10 structured review questions with hints, common mistakes, and concepts-at-a-glance tables |
| Advanced Perspectives | 5 Theory Connection slides (AI/ML applications) + 5 PhD View slides (research-level topics) |
Part I: Mathematical Foundations¶
Lecture 0: Introduction & Course Overview¶
Course organization, assessment structure, textbooks, teaching philosophy, notation conventions, and what to expect.
Lecture 2: Linear Algebra¶
Foundation of Everything in AI
Linear algebra is the computational engine behind all of machine learning. Every dataset is a matrix, every data point is a vector, and every neural network layer computes a linear transformation.
Topics covered (8 sections):
- Systems of Linear Equations
- Matrices
- Solving Systems of Linear Equations
- Vector Spaces
- Linear Independence
- Basis and Rank
- Linear Mappings
- Affine Spaces
What's inside the slides:
- Definitions with plain-English "Think of it as..." intuitions
- Full Math Tutorial section with 15 worked examples
- 10 Review Questions with hints (systems, matrices, inverse/transpose, Gaussian elimination, vector spaces, linear independence, basis/dimension, rank, linear mappings, affine spaces)
- Common Mistakes to Avoid (5 key pitfalls)
- Concepts at a Glance comparison table
- Key Takeaways summary
- 5 Theory Connection slides (ML, Computer Vision, Deep Learning, Optimization, Data Science)
- 5 PhD View slides (Functional Analysis, Matrix Decompositions, Tensor Algebra, Numerical LA, Spectral Theory)
Materials:
Lecture 3: Analytic Geometry¶
The Geometry Behind Machine Learning
Analytic geometry gives us the tools to measure similarity, distance, and angles in high-dimensional spaces — the foundation of recommendation systems, NLP embeddings, and PCA.
Topics covered (9 sections):
- Norms
- Inner Products
- Lengths and Distances
- Angles and Orthogonality
- Orthonormal Basis
- Orthogonal Complement
- Inner Product of Functions
- Orthogonal Projections
- Rotations
What's inside the slides:
- Complete concept chain: Inner Product → Norm → Distance → Angle → Orthogonality → Projection
- Projection formulas for 1D lines and general subspaces
- Gram-Schmidt orthogonalization with worked examples
- 10 Review Questions with hints (norms, inner products, Cauchy-Schwarz, angles, orthogonal matrices, Gram-Schmidt, orthogonal complement, projections onto lines, general projections, rotations)
- Common Mistakes to Avoid (5 key pitfalls)
- Concepts at a Glance comparison table
- 5 Theory Connection slides (ML, NLP, Deep Learning, Computer Vision, Data Science)
- 5 PhD View slides (Hilbert Spaces, RKHS, Compressed Sensing, Riemannian Geometry, Random Projections)
Materials:
- Download Lecture 3: Analytic Geometry (PDF)
- Tutorial: Analytic Geometry
- Notebook: Analytic Geometry in Python
Lecture 4: Matrix Decomposition¶
Revealing Hidden Structure in Data
Matrix decompositions like eigendecomposition and SVD are the workhorses of data science — they power PCA, recommender systems, image compression, and Google's PageRank.
Topics covered (6 sections):
- Determinants and Trace
- Eigenvalues and Eigenvectors
- Cholesky Decomposition
- Eigendecomposition and Diagonalization
- Singular Value Decomposition (SVD)
- Matrix Approximation
What's inside the slides:
- Eigenvalue computation with characteristic polynomials
- SVD: geometric interpretation and step-by-step examples
- Matrix approximation via truncated SVD
- 10 Review Questions with hints
- Common Mistakes to Avoid
- 5 Theory Connection slides (ML applications of decompositions)
- 5 PhD View slides (advanced decomposition theory)
Materials:
- Download Lecture 4: Matrix Decomposition (PDF)
- Tutorial: Matrix Decomposition
- Notebook: Matrix Decomposition in Python
Lecture 5: Vector Calculus¶
The Mathematics of Learning
Vector calculus provides the gradient and chain rule — without these, there is no backpropagation, no gradient descent, and no training of neural networks.
Topics covered (6 sections):
- Differentiation of Univariate Functions
- Partial Differentiation and Gradients
- Gradients of Vector-Valued Functions
- Gradients of Matrices
- Useful Identities for Computing Gradients
- Backpropagation and Automatic Differentiation
What's inside the slides:
- From single-variable derivatives to Jacobians and gradients
- Chain rule for vector and matrix functions
- Backpropagation derivation with computation graphs
- 10 Review Questions with hints
- Common Mistakes to Avoid
- 5 Theory Connection slides (optimization, neural networks, physics)
- 5 PhD View slides (differential geometry, automatic differentiation theory)
Materials:
- Download Lecture 5: Vector Calculus (PDF)
- Tutorial: Vector Calculus
- Notebook: Vector Calculus in Python
Lecture 6: Probability & Distributions¶
Reasoning Under Uncertainty
Probability theory is the language of uncertainty — it underpins Bayesian inference, generative models, statistical testing, and every probabilistic ML algorithm.
Topics covered (8 sections):
- Construction of a Probability Space
- Discrete and Continuous Probabilities
- Sum Rule, Product Rule, and Bayes' Theorem
- Summary Statistics and Independence
- Gaussian Distribution
- Conjugacy and Exponential Family
- Change of Variables / Inverse Transform
- Probability in Machine Learning
What's inside the slides:
- From sample spaces to random variables to distributions
- Discrete (Bernoulli, Binomial, Geometric) and Continuous (Uniform, Exponential, Gaussian) distributions
- Bayes' Theorem with real-world examples
- Marginal and conditional distributions
- Covariance, correlation, and independence
- 10 Review Questions with hints
- Common Mistakes to Avoid
- 5 Theory Connection slides (Bayesian ML, generative models, information theory)
- 5 PhD View slides (measure theory, stochastic processes, information geometry)
Materials:
- Download Lecture 6: Probability & Distributions (PDF)
- Tutorial: Probability and Distributions
- Notebook: Probability and Distributions in Python
Part II: Machine Learning Applications¶
Application 1: When Models Meet Data¶
Description: Introduction to the practical aspects of applying mathematical models to real-world datasets and the challenges that arise.
What you'll learn:
- Data preprocessing and feature engineering
- Model selection and evaluation
- Overfitting and regularization
- Bias-variance tradeoff
Materials:
Application 2: Dimensionality Reduction (PCA)¶
Description: Using Principal Component Analysis to reduce data complexity while preserving essential information for efficient learning.
What you'll learn:
- PCA theory: eigenvalues meet data variance
- Step-by-step PCA computation
- Variance explained and choosing dimensions
- Eigenfaces and image compression
Materials:
Application 3: Density Estimation (GMM)¶
Description: Probabilistic approaches to understanding data distributions and clustering using Gaussian Mixture Models.
What you'll learn:
- Probability density estimation
- Gaussian Mixture Models
- Expectation-Maximization (EM) algorithm
- Clustering applications
Materials:
Application 4: Classification (SVM)¶
Description: Geometric and optimization-based methods for supervised learning and decision boundary determination using Support Vector Machines.
What you'll learn:
- Support Vector Machines (SVM)
- Kernel methods and the kernel trick
- Margin maximization
- Multi-class classification
Materials:
Quick Reference: All Lecture Downloads¶
| Lecture | Topic | Download |
|---|---|---|
| Lecture 0 | Introduction & Course Overview | |
| Lecture 2 | Linear Algebra | |
| Lecture 3 | Analytic Geometry | |
| Lecture 4 | Matrix Decomposition | |
| Lecture 5 | Vector Calculus | |
| Lecture 6 | Probability & Distributions | |
| Applications | ML Applications (all 4) |
Primary Textbooks¶
Mathematics for Machine Learning¶
Authors: Deisenroth, Faisal, and Ong Publisher: Cambridge University Press Website: mml-book.github.io
Convex Optimization¶
Authors: Boyd and Vandenberghe Publisher: Cambridge University Press Website: stanford.edu/~boyd/cvxbook
Introduction to Probability¶
Authors: Bertsekas and Tsitsiklis Publisher: Athena Scientific (2nd Ed.)
Learning Tips¶
Before Each Lecture:
- Download and review the lecture slides
- Complete prerequisite readings
- Review previous lecture concepts
During Lectures:
- Follow the four-part structure
- Take notes during explanations
- Work through examples actively
- Try coding exercises immediately
After Lectures:
- Work through the 10 Review Questions (use hints if stuck)
- Study the Common Mistakes to Avoid
- Explore Theory Connection slides for AI/ML context
- Review PhD View slides for deeper understanding
- Complete Python notebooks
Lectures prepared by Mohammed Alnemari Mathematics of AI • Spring 2026
Last Updated: February 8, 2026