Lesson overview | Previous part | Next part
Hyperparameter Optimization: Part 5: Core Theory III: Practical Variants to 6. Advanced Topics
5. Core Theory III: Practical Variants
This block develops core theory iii: practical variants for Hyperparameter Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
5.1 Variant built around upper confidence bound
In this section, population-based training is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Variant built around upper confidence bound" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, population-based training is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where population-based training can be computed directly and compared with theory.
- A logistic-regression or softmax objective where population-based training affects optimization but the model remains interpretable.
- A transformer training diagnostic where population-based training appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating population-based training as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving population-based training, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes population-based training visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about population-based training is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.2 Variant built around Thompson sampling
In this section, multi-objective tuning is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Variant built around Thompson sampling" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, multi-objective tuning is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where multi-objective tuning can be computed directly and compared with theory.
- A logistic-regression or softmax objective where multi-objective tuning affects optimization but the model remains interpretable.
- A transformer training diagnostic where multi-objective tuning appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating multi-objective tuning as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving multi-objective tuning, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes multi-objective tuning visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about multi-objective tuning is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.3 Variant built around Bayesian optimization
In this section, Pareto frontier is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Variant built around Bayesian optimization" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Pareto frontier is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Pareto frontier can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Pareto frontier affects optimization but the model remains interpretable.
- A transformer training diagnostic where Pareto frontier appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Pareto frontier as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Pareto frontier, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Pareto frontier visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Pareto frontier is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.4 Implementation constraints and numerical stability
In this section, validation leakage is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Implementation constraints and numerical stability" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, validation leakage is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where validation leakage can be computed directly and compared with theory.
- A logistic-regression or softmax objective where validation leakage affects optimization but the model remains interpretable.
- A transformer training diagnostic where validation leakage appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating validation leakage as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving validation leakage, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes validation leakage visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about validation leakage is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.5 What belongs here versus neighboring sections
In this section, nested validation is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "What belongs here versus neighboring sections" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, nested validation is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where nested validation can be computed directly and compared with theory.
- A logistic-regression or softmax objective where nested validation affects optimization but the model remains interpretable.
- A transformer training diagnostic where nested validation appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating nested validation as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving nested validation, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes nested validation visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about nested validation is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6. Advanced Topics
This block develops advanced topics for Hyperparameter Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
6.1 Advanced view of successive halving
In this section, validation leakage is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Advanced view of successive halving" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, validation leakage is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where validation leakage can be computed directly and compared with theory.
- A logistic-regression or softmax objective where validation leakage affects optimization but the model remains interpretable.
- A transformer training diagnostic where validation leakage appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating validation leakage as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving validation leakage, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes validation leakage visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about validation leakage is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.2 Advanced view of Hyperband
In this section, nested validation is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Advanced view of Hyperband" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, nested validation is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where nested validation can be computed directly and compared with theory.
- A logistic-regression or softmax objective where nested validation affects optimization but the model remains interpretable.
- A transformer training diagnostic where nested validation appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating nested validation as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving nested validation, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes nested validation visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about nested validation is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.3 Advanced view of ASHA
In this section, LLM fine-tuning search is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Advanced view of ASHA" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, LLM fine-tuning search is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where LLM fine-tuning search can be computed directly and compared with theory.
- A logistic-regression or softmax objective where LLM fine-tuning search affects optimization but the model remains interpretable.
- A transformer training diagnostic where LLM fine-tuning search appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating LLM fine-tuning search as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving LLM fine-tuning search, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes LLM fine-tuning search visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about LLM fine-tuning search is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.4 Infinite-dimensional or large-scale interpretation
In this section, configuration space is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Infinite-dimensional or large-scale interpretation" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, configuration space is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where configuration space can be computed directly and compared with theory.
- A logistic-regression or softmax objective where configuration space affects optimization but the model remains interpretable.
- A transformer training diagnostic where configuration space appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating configuration space as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving configuration space, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes configuration space visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about configuration space is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.5 Open questions for frontier model training
In this section, conditional parameter is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Open questions for frontier model training" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, conditional parameter is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where conditional parameter can be computed directly and compared with theory.
- A logistic-regression or softmax objective where conditional parameter affects optimization but the model remains interpretable.
- A transformer training diagnostic where conditional parameter appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating conditional parameter as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving conditional parameter, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes conditional parameter visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about conditional parameter is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.