Foundations Track¶
This track builds core ML execution discipline: math intuition, data understanding, feature preparation, robust splitting, and evaluation quality.
Recommended Sequence¶
projects/deep-learning-math-foundations-showcaseprojects/neural-network-foundations-showcaseprojects/pytorch-training-regularization-showcaseprojects/sota-supervised-learning-showcaseprojects/eda-leakage-profiling-showcaseprojects/feature-engineering-dimred-showcaseprojects/sota-unsupervised-semisup-showcase
Core Skills Covered¶
- Vectors, matrices, gradients, entropy, and optimization traces.
- Perceptrons, activations, backpropagation, and initialization choices.
- PyTorch tensors, training loops, optimizers, schedulers, and regularization tradeoffs.
- Train/validation/test discipline.
- Stratified, group-aware, and time-aware split strategies.
- Univariate and bivariate EDA.
- Missingness diagnostics and leakage checks.
- Feature encoding, selection, and dimensionality reduction.
Evidence Artifacts To Inspect¶
artifacts/vector_operations.csvartifacts/gradient_descent_trace.csvartifacts/activation_comparison.csvartifacts/decision_boundary_summary.csvartifacts/optimizer_comparison.csvartifacts/regularization_ablation.csvartifacts/splits/split_manifest.jsonartifacts/eda/univariate_summary.csvartifacts/eda/bivariate_vs_target.csvartifacts/leakage/leakage_report.csvartifacts/selection/selection_scores.csv
Suggested Reflection Prompts¶
- Which split strategy is most defensible for this dataset and why?
- Where does model capacity start to help, and where does it start to overfit?
- Which optimizer or regularizer improved learning without hiding what the model was doing?
- Which feature engineering step improved generalization most?
- Which leakage check would fail first if pipeline order changed?