Lesson overview | Previous part | Next part
Gradient Descent: Part 3: Core Theory I: Geometry and Guarantees to 4. Core Theory II: Algorithms and Dynamics
3. Core Theory I: Geometry and Guarantees
This block develops core theory i: geometry and guarantees for Gradient Descent. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
3.1 Geometry of exact line search
In this section, PL condition is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Geometry of exact line search" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, PL condition is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where PL condition can be computed directly and compared with theory.
- A logistic-regression or softmax objective where PL condition affects optimization but the model remains interpretable.
- A transformer training diagnostic where PL condition appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating PL condition as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving PL condition, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes PL condition visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about PL condition is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.2 Key inequality for backtracking line search
In this section, condition number is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Key inequality for backtracking line search" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, condition number is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where condition number can be computed directly and compared with theory.
- A logistic-regression or softmax objective where condition number affects optimization but the model remains interpretable.
- A transformer training diagnostic where condition number appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating condition number as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving condition number, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes condition number visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about condition number is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.3 Role of Armijo condition
In this section, Polyak momentum is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Role of Armijo condition" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Polyak momentum is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Polyak momentum can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Polyak momentum affects optimization but the model remains interpretable.
- A transformer training diagnostic where Polyak momentum appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Polyak momentum as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Polyak momentum, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Polyak momentum visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Polyak momentum is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.4 Proof template and what the proof actually buys
In this section, Nesterov acceleration is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Proof template and what the proof actually buys" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Nesterov acceleration is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Nesterov acceleration can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Nesterov acceleration affects optimization but the model remains interpretable.
- A transformer training diagnostic where Nesterov acceleration appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Nesterov acceleration as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Nesterov acceleration, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Nesterov acceleration visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Nesterov acceleration is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.5 Failure modes when assumptions are removed
In this section, gradient flow is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Failure modes when assumptions are removed" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, gradient flow is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where gradient flow can be computed directly and compared with theory.
- A logistic-regression or softmax objective where gradient flow affects optimization but the model remains interpretable.
- A transformer training diagnostic where gradient flow appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating gradient flow as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving gradient flow, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes gradient flow visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about gradient flow is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4. Core Theory II: Algorithms and Dynamics
This block develops core theory ii: algorithms and dynamics for Gradient Descent. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
4.1 Algorithmic update for Wolfe conditions
In this section, Nesterov acceleration is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Algorithmic update for Wolfe conditions" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Nesterov acceleration is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Nesterov acceleration can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Nesterov acceleration affects optimization but the model remains interpretable.
- A transformer training diagnostic where Nesterov acceleration appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Nesterov acceleration as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Nesterov acceleration, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Nesterov acceleration visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Nesterov acceleration is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.2 Stability role of convex convergence
In this section, gradient flow is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Stability role of convex convergence" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, gradient flow is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where gradient flow can be computed directly and compared with theory.
- A logistic-regression or softmax objective where gradient flow affects optimization but the model remains interpretable.
- A transformer training diagnostic where gradient flow appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating gradient flow as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving gradient flow, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes gradient flow visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about gradient flow is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.3 Rate or complexity controlled by strongly convex convergence
In this section, oscillation is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Rate or complexity controlled by strongly convex convergence" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, oscillation is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where oscillation can be computed directly and compared with theory.
- A logistic-regression or softmax objective where oscillation affects optimization but the model remains interpretable.
- A transformer training diagnostic where oscillation appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating oscillation as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving oscillation, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes oscillation visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about oscillation is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.4 Diagnostic interpretation of the update path
In this section, edge of stability preview is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Diagnostic interpretation of the update path" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, edge of stability preview is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where edge of stability preview can be computed directly and compared with theory.
- A logistic-regression or softmax objective where edge of stability preview affects optimization but the model remains interpretable.
- A transformer training diagnostic where edge of stability preview appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating edge of stability preview as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving edge of stability preview, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes edge of stability preview visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about edge of stability preview is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.5 Connection to the next section in the chapter
In this section, gradient clipping preview is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Gradient Descent, the phrase "Connection to the next section in the chapter" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, gradient clipping preview is the part of Gradient Descent that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where gradient clipping preview can be computed directly and compared with theory.
- A logistic-regression or softmax objective where gradient clipping preview affects optimization but the model remains interpretable.
- A transformer training diagnostic where gradient clipping preview appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating gradient clipping preview as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving gradient clipping preview, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes gradient clipping preview visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about gradient clipping preview is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- the basic training loop used by every neural-network optimizer.
- step-size stability for cross-entropy and mean-squared-error objectives.
- momentum as the ancestor of Adam's first-moment accumulator.
- line-search logic as a debugging model for divergence and oscillation.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.