Lesson overview | Previous part | Lesson overview
Regularization Methods: Part 7: Applications in Machine Learning to Appendix A. Extended Derivation and Diagnostic Cards
7. Applications in Machine Learning
This block develops applications in machine learning for Regularization Methods. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
7.1 weight decay in AdamW-based transformer training
In this section, explicit penalty is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "weight decay in AdamW-based transformer training" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, explicit penalty is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where explicit penalty can be computed directly and compared with theory.
- A logistic-regression or softmax objective where explicit penalty affects optimization but the model remains interpretable.
- A transformer training diagnostic where explicit penalty appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating explicit penalty as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving explicit penalty, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes explicit penalty visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about explicit penalty is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
7.2 dropout and stochastic regularization for neural networks
In this section, constraint equivalence is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "dropout and stochastic regularization for neural networks" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, constraint equivalence is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where constraint equivalence can be computed directly and compared with theory.
- A logistic-regression or softmax objective where constraint equivalence affects optimization but the model remains interpretable.
- A transformer training diagnostic where constraint equivalence appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating constraint equivalence as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving constraint equivalence, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes constraint equivalence visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about constraint equivalence is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
7.3 spectral normalization in GANs and Lipschitz-controlled models
In this section, L2 penalty is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "spectral normalization in GANs and Lipschitz-controlled models" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, L2 penalty is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where L2 penalty can be computed directly and compared with theory.
- A logistic-regression or softmax objective where L2 penalty affects optimization but the model remains interpretable.
- A transformer training diagnostic where L2 penalty appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating L2 penalty as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving L2 penalty, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes L2 penalty visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about L2 penalty is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
7.4 SAM as a regularizer that penalizes sharp local neighborhoods
In this section, weight decay is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "SAM as a regularizer that penalizes sharp local neighborhoods" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, weight decay is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where weight decay can be computed directly and compared with theory.
- A logistic-regression or softmax objective where weight decay affects optimization but the model remains interpretable.
- A transformer training diagnostic where weight decay appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating weight decay as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving weight decay, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes weight decay visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about weight decay is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
7.5 Diagnostic checklist for real experiments
In this section, AdamW decay is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Diagnostic checklist for real experiments" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, AdamW decay is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where AdamW decay can be computed directly and compared with theory.
- A logistic-regression or softmax objective where AdamW decay affects optimization but the model remains interpretable.
- A transformer training diagnostic where AdamW decay appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating AdamW decay as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving AdamW decay, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes AdamW decay visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about AdamW decay is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
8. Implementation and Diagnostics
This block develops implementation and diagnostics for Regularization Methods. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
8.1 Minimal NumPy experiment for SAM
In this section, weight decay is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Minimal NumPy experiment for SAM" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, weight decay is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where weight decay can be computed directly and compared with theory.
- A logistic-regression or softmax objective where weight decay affects optimization but the model remains interpretable.
- A transformer training diagnostic where weight decay appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating weight decay as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving weight decay, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes weight decay visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about weight decay is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
8.2 Monitoring signal for implicit regularization
In this section, AdamW decay is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Monitoring signal for implicit regularization" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, AdamW decay is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where AdamW decay can be computed directly and compared with theory.
- A logistic-regression or softmax objective where AdamW decay affects optimization but the model remains interpretable.
- A transformer training diagnostic where AdamW decay appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating AdamW decay as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving AdamW decay, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes AdamW decay visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about AdamW decay is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
8.3 Failure signature for Bayesian MAP view
In this section, L1 penalty is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Failure signature for Bayesian MAP view" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, L1 penalty is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where L1 penalty can be computed directly and compared with theory.
- A logistic-regression or softmax objective where L1 penalty affects optimization but the model remains interpretable.
- A transformer training diagnostic where L1 penalty appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating L1 penalty as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving L1 penalty, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes L1 penalty visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about L1 penalty is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
8.4 Framework-level implementation pattern
In this section, soft thresholding is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Framework-level implementation pattern" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, soft thresholding is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where soft thresholding can be computed directly and compared with theory.
- A logistic-regression or softmax objective where soft thresholding affects optimization but the model remains interpretable.
- A transformer training diagnostic where soft thresholding appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating soft thresholding as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving soft thresholding, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes soft thresholding visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about soft thresholding is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
8.5 Reproducibility and logging checklist
In this section, elastic net is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Regularization Methods, the phrase "Reproducibility and logging checklist" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, elastic net is the part of Regularization Methods that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where elastic net can be computed directly and compared with theory.
- A logistic-regression or softmax objective where elastic net affects optimization but the model remains interpretable.
- A transformer training diagnostic where elastic net appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating elastic net as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving elastic net, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes elastic net visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about elastic net is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- weight decay in AdamW-based transformer training.
- dropout and stochastic regularization for neural networks.
- spectral normalization in GANs and Lipschitz-controlled models.
- SAM as a regularizer that penalizes sharp local neighborhoods.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
9. Common Mistakes
| # | Mistake | Why It Is Wrong | Fix |
|---|---|---|---|
| 1 | Using a recipe without checking assumptions | Optimization guarantees depend on smoothness, convexity, stochasticity, or feasibility assumptions. | Write the assumptions next to the update rule before choosing hyperparameters. |
| 2 | Confusing objective decrease with validation improvement | The optimizer sees the training objective; validation behavior also depends on generalization and data split quality. | Track objective, train metric, validation metric, and update norm separately. |
| 3 | Treating all norms as interchangeable | The geometry changes when the norm changes, especially for constraints and regularizers. | State whether you use , , Frobenius, spectral, or another norm. |
| 4 | Ignoring scale | Learning rates, penalties, curvature, and gradient norms are all scale-sensitive. | Normalize units and inspect effective update size . |
| 5 | Overfitting to a single seed | Optimization can look stable for one seed and fail under another. | Run small seed sweeps for important claims. |
| 6 | Hiding instability behind smoothed plots | A moving average can hide spikes, divergence, and bad curvature events. | Plot raw metrics alongside smoothed metrics. |
| 7 | Using test data during tuning | This contaminates the final evaluation. | Reserve test data until after model and hyperparameter selection. |
| 8 | Assuming large models make theory irrelevant | Large models often make diagnostics more important because failures are expensive. | Use theory to decide what to log, not to pretend every theorem applies exactly. |
| 9 | Mixing optimizer state with model state carelessly | State corruption changes the effective algorithm. | Checkpoint parameters, gradients if needed, optimizer moments, scheduler state, and random seeds. |
| 10 | Not checking numerical precision | BF16, FP16, FP8, and accumulation choices can change the observed optimizer. | Cross-check suspicious runs against higher precision on a small batch. |
10. Exercises
- Exercise 1 [*] - L2 Penalty (a) Define L2 penalty using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 2 [*] - Adamw Decay (a) Define AdamW decay using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 3 [*] - Soft Thresholding (a) Define soft thresholding using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 4 [] - Nuclear Norm** (a) Define nuclear norm using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 5 [] - Early Stopping** (a) Define early stopping using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 6 [] - Label Smoothing Preview** (a) Define label smoothing preview using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 7 [] - Gradient Clipping Preview** (a) Define gradient clipping preview using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 8 [*] - Implicit Regularization** (a) Define implicit regularization using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 9 [*] - Double Descent** (a) Define double descent using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
- Exercise 10 [*] - Lora Rank Regularity** (a) Define LoRA rank regularity using the notation of this repository. (b) Give three valid examples and two non-examples. (c) Derive the relevant update or inequality shown below.
(d) Implement a NumPy check on a synthetic two-dimensional objective. (e) Explain what metric you would log in a real LLM or fine-tuning run.
11. Why This Matters for AI (2026 Perspective)
| Concept | AI Impact |
|---|---|
| explicit penalty | weight decay in AdamW-based transformer training |
| constraint equivalence | dropout and stochastic regularization for neural networks |
| L2 penalty | spectral normalization in GANs and Lipschitz-controlled models |
| weight decay | SAM as a regularizer that penalizes sharp local neighborhoods |
| AdamW decay | weight decay in AdamW-based transformer training |
| L1 penalty | dropout and stochastic regularization for neural networks |
| soft thresholding | spectral normalization in GANs and Lipschitz-controlled models |
| elastic net | SAM as a regularizer that penalizes sharp local neighborhoods |
| nuclear norm | weight decay in AdamW-based transformer training |
| dropout | dropout and stochastic regularization for neural networks |
12. Conceptual Bridge
Regularization Methods sits inside a chain. Earlier sections give the calculus, probability, and linear algebra needed to write the objective and interpret the update. Later sections use this material to reason about noisy gradients, adaptive state, regularization, tuning, schedules, and finally information-theoretic losses.
Backward link: Adaptive Learning Rate supplies the immediate prerequisite vocabulary.
Forward link: Hyperparameter Optimization uses this section as a building block.
+------------------------------------------------------------+
| Chapter 8: Optimization |
| 01-Convex-Optimization Convex Optimization |
| 02-Gradient-Descent Gradient Descent |
| 03-Second-Order-Methods Second-Order Methods |
| 04-Constrained-Optimization Constrained Optimization |
| 05-Stochastic-Optimization Stochastic Optimization |
| 06-Optimization-Landscape Optimization Landscape |
| 07-Adaptive-Learning-Rate Adaptive Learning Rate |
| >> 08-Regularization-Methods Regularization Methods |
| 09-Hyperparameter-Optimization Hyperparameter Optimization |
| 10-Learning-Rate-Schedules Learning Rate Schedules |
+------------------------------------------------------------+
Appendix A. Extended Derivation and Diagnostic Cards
References
- Tibshirani, Regression Shrinkage and Selection via the Lasso.
- Srivastava et al., Dropout.
- Loshchilov and Hutter, Decoupled Weight Decay Regularization.
- Miyato et al., Spectral Normalization for GANs.
- Foret et al., Sharpness-Aware Minimization.
- Goodfellow, Bengio, and Courville, Deep Learning.
- Bottou, Curtis, and Nocedal, Optimization Methods for Large-Scale Machine Learning.
- PyTorch optimizer and scheduler documentation.
- Optax documentation for composable optimizer transformations.