Lesson overview | Previous part | Next part
Gradient Descent: Part 5: Core Theory III: Practical Variants to 6. Advanced Topics
5. Core Theory III: Practical Variants
This block develops core theory iii: practical variants for Gradient Descent. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
5.1 Variant built around nonconvex stationarity
In this section, edge of stability preview is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Variant built around nonconvex stationarity" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, edge of stability preview is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where edge of stability preview can be computed directly and compared with theory.
- A logistic-regression or softmax objective where edge of stability preview affects optimization but the model remains interpretable.
- A transformer training diagnostic where edge of stability preview appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating edge of stability preview as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving edge of stability preview, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes edge of stability preview visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about edge of stability preview is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.2 Variant built around PL condition
In this section, gradient clipping preview is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Variant built around PL condition" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, gradient clipping preview is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where gradient clipping preview can be computed directly and compared with theory.
- A logistic-regression or softmax objective where gradient clipping preview affects optimization but the model remains interpretable.
- A transformer training diagnostic where gradient clipping preview appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating gradient clipping preview as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving gradient clipping preview, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes gradient clipping preview visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about gradient clipping preview is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.3 Variant built around condition number
In this section, linear regression by GD is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Variant built around condition number" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, linear regression by GD is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where linear regression by GD can be computed directly and compared with theory.
- A logistic-regression or softmax objective where linear regression by GD affects optimization but the model remains interpretable.
- A transformer training diagnostic where linear regression by GD appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating linear regression by GD as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving linear regression by GD, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes linear regression by GD visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about linear regression by GD is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.4 Implementation constraints and numerical stability
In this section, logistic regression by GD is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Implementation constraints and numerical stability" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, logistic regression by GD is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where logistic regression by GD can be computed directly and compared with theory.
- A logistic-regression or softmax objective where logistic regression by GD affects optimization but the model remains interpretable.
- A transformer training diagnostic where logistic regression by GD appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating logistic regression by GD as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving logistic regression by GD, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes logistic regression by GD visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about logistic regression by GD is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.5 What belongs here versus neighboring sections
In this section, learning-rate diagnostics is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "What belongs here versus neighboring sections" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, learning-rate diagnostics is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where learning-rate diagnostics can be computed directly and compared with theory.
- A logistic-regression or softmax objective where learning-rate diagnostics affects optimization but the model remains interpretable.
- A transformer training diagnostic where learning-rate diagnostics appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating learning-rate diagnostics as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving learning-rate diagnostics, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes learning-rate diagnostics visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about learning-rate diagnostics is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6. Advanced Topics
This block develops advanced topics for Gradient Descent. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
6.1 Advanced view of Polyak momentum
In this section, logistic regression by GD is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Advanced view of Polyak momentum" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, logistic regression by GD is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where logistic regression by GD can be computed directly and compared with theory.
- A logistic-regression or softmax objective where logistic regression by GD affects optimization but the model remains interpretable.
- A transformer training diagnostic where logistic regression by GD appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating logistic regression by GD as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving logistic regression by GD, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes logistic regression by GD visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about logistic regression by GD is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.2 Advanced view of Nesterov acceleration
In this section, learning-rate diagnostics is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Advanced view of Nesterov acceleration" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, learning-rate diagnostics is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where learning-rate diagnostics can be computed directly and compared with theory.
- A logistic-regression or softmax objective where learning-rate diagnostics affects optimization but the model remains interpretable.
- A transformer training diagnostic where learning-rate diagnostics appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating learning-rate diagnostics as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving learning-rate diagnostics, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes learning-rate diagnostics visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about learning-rate diagnostics is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.3 Advanced view of gradient flow
In this section, optimization loop design is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Advanced view of gradient flow" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, optimization loop design is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where optimization loop design can be computed directly and compared with theory.
- A logistic-regression or softmax objective where optimization loop design affects optimization but the model remains interpretable.
- A transformer training diagnostic where optimization loop design appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating optimization loop design as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving optimization loop design, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes optimization loop design visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about optimization loop design is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.4 Infinite-dimensional or large-scale interpretation
In this section, gradient direction is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Infinite-dimensional or large-scale interpretation" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, gradient direction is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where gradient direction can be computed directly and compared with theory.
- A logistic-regression or softmax objective where gradient direction affects optimization but the model remains interpretable.
- A transformer training diagnostic where gradient direction appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating gradient direction as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving gradient direction, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes gradient direction visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about gradient direction is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.5 Open questions for frontier model training
In this section, descent lemma is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Open questions for frontier model training" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, descent lemma is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where descent lemma can be computed directly and compared with theory.
- A logistic-regression or softmax objective where descent lemma affects optimization but the model remains interpretable.
- A transformer training diagnostic where descent lemma appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating descent lemma as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving descent lemma, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes descent lemma visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about descent lemma is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.