Lesson overview | Lesson overview | Next part
Adaptive Learning Rate: Part 1: Intuition to 2. Formal Definitions
1. Intuition
This block develops intuition for Adaptive Learning Rate. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
1.1 Why Adaptive Learning Rate matters for training systems
In this section, Adam first moment is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Why Adaptive Learning Rate matters for training systems" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Adam first moment is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Adam first moment can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Adam first moment affects optimization but the model remains interpretable.
- A transformer training diagnostic where Adam first moment appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Adam first moment as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Adam first moment, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Adam first moment visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Adam first moment is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
1.2 The optimization object: parameters, objective, algorithm, and diagnostic
In this section, Adam second moment is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "The optimization object: parameters, objective, algorithm, and diagnostic" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Adam second moment is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Adam second moment can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Adam second moment affects optimization but the model remains interpretable.
- A transformer training diagnostic where Adam second moment appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Adam second moment as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Adam second moment, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Adam second moment visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Adam second moment is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
1.3 Historical arc from classical optimization to modern AI
In this section, bias correction is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Historical arc from classical optimization to modern AI" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, bias correction is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where bias correction can be computed directly and compared with theory.
- A logistic-regression or softmax objective where bias correction affects optimization but the model remains interpretable.
- A transformer training diagnostic where bias correction appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating bias correction as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving bias correction, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes bias correction visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about bias correction is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
1.4 What this section treats as canonical scope
In this section, epsilon stabilizer is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "What this section treats as canonical scope" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, epsilon stabilizer is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where epsilon stabilizer can be computed directly and compared with theory.
- A logistic-regression or softmax objective where epsilon stabilizer affects optimization but the model remains interpretable.
- A transformer training diagnostic where epsilon stabilizer appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating epsilon stabilizer as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving epsilon stabilizer, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes epsilon stabilizer visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about epsilon stabilizer is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
1.5 A first mental model for LLM training
In this section, AMSGrad is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "A first mental model for LLM training" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, AMSGrad is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where AMSGrad can be computed directly and compared with theory.
- A logistic-regression or softmax objective where AMSGrad affects optimization but the model remains interpretable.
- A transformer training diagnostic where AMSGrad appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating AMSGrad as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving AMSGrad, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes AMSGrad visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about AMSGrad is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
2. Formal Definitions
This block develops formal definitions for Adaptive Learning Rate. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
2.1 Primary definition: effective learning rate
In this section, epsilon stabilizer is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Primary definition: effective learning rate" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, epsilon stabilizer is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where epsilon stabilizer can be computed directly and compared with theory.
- A logistic-regression or softmax objective where epsilon stabilizer affects optimization but the model remains interpretable.
- A transformer training diagnostic where epsilon stabilizer appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating epsilon stabilizer as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving epsilon stabilizer, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes epsilon stabilizer visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about epsilon stabilizer is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
2.2 Secondary definition: diagonal preconditioner
In this section, AMSGrad is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Secondary definition: diagonal preconditioner" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, AMSGrad is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where AMSGrad can be computed directly and compared with theory.
- A logistic-regression or softmax objective where AMSGrad affects optimization but the model remains interpretable.
- A transformer training diagnostic where AMSGrad appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating AMSGrad as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving AMSGrad, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes AMSGrad visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about AMSGrad is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
2.3 Algorithmic object: AdaGrad accumulator
In this section, AdamW is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Algorithmic object: AdaGrad accumulator" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, AdamW is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where AdamW can be computed directly and compared with theory.
- A logistic-regression or softmax objective where AdamW affects optimization but the model remains interpretable.
- A transformer training diagnostic where AdamW appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating AdamW as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving AdamW, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes AdamW visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about AdamW is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
2.4 Examples, non-examples, and boundary cases
In this section, coupled L2 is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Examples, non-examples, and boundary cases" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, coupled L2 is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where coupled L2 can be computed directly and compared with theory.
- A logistic-regression or softmax objective where coupled L2 affects optimization but the model remains interpretable.
- A transformer training diagnostic where coupled L2 appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating coupled L2 as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving coupled L2, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes coupled L2 visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about coupled L2 is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
2.5 Notation, dimensions, and assumptions
In this section, decoupled weight decay is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Adaptive Learning Rate, the phrase "Notation, dimensions, and assumptions" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, decoupled weight decay is the part of Adaptive Learning Rate that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where decoupled weight decay can be computed directly and compared with theory.
- A logistic-regression or softmax objective where decoupled weight decay affects optimization but the model remains interpretable.
- A transformer training diagnostic where decoupled weight decay appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating decoupled weight decay as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving decoupled weight decay, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes decoupled weight decay visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about decoupled weight decay is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- AdamW as the default optimizer for transformer pretraining and fine-tuning.
- Adafactor for memory-constrained large models.
- LAMB and LARS for large-batch training.
- optimizer-state diagnostics for training failures and loss spikes.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.