Lesson overview | Previous part | Next part
Hyperparameter Optimization: Part 3: Core Theory I: Geometry and Guarantees to 4. Core Theory II: Algorithms and Dynamics
3. Core Theory I: Geometry and Guarantees
This block develops core theory i: geometry and guarantees for Hyperparameter Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
3.1 Geometry of grid search
In this section, Thompson sampling is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Geometry of grid search" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Thompson sampling is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Thompson sampling can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Thompson sampling affects optimization but the model remains interpretable.
- A transformer training diagnostic where Thompson sampling appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Thompson sampling as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Thompson sampling, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Thompson sampling visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Thompson sampling is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.2 Key inequality for random search
In this section, Bayesian optimization is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Key inequality for random search" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Bayesian optimization is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Bayesian optimization can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Bayesian optimization affects optimization but the model remains interpretable.
- A transformer training diagnostic where Bayesian optimization appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Bayesian optimization as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Bayesian optimization, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Bayesian optimization visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Bayesian optimization is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.3 Role of Sobol initialization
In this section, successive halving is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Role of Sobol initialization" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, successive halving is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where successive halving can be computed directly and compared with theory.
- A logistic-regression or softmax objective where successive halving affects optimization but the model remains interpretable.
- A transformer training diagnostic where successive halving appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating successive halving as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving successive halving, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes successive halving visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about successive halving is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.4 Proof template and what the proof actually buys
In this section, Hyperband is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Proof template and what the proof actually buys" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Hyperband is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Hyperband can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Hyperband affects optimization but the model remains interpretable.
- A transformer training diagnostic where Hyperband appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Hyperband as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Hyperband, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Hyperband visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Hyperband is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.5 Failure modes when assumptions are removed
In this section, ASHA is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Failure modes when assumptions are removed" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, ASHA is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where ASHA can be computed directly and compared with theory.
- A logistic-regression or softmax objective where ASHA affects optimization but the model remains interpretable.
- A transformer training diagnostic where ASHA appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating ASHA as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving ASHA, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes ASHA visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about ASHA is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4. Core Theory II: Algorithms and Dynamics
This block develops core theory ii: algorithms and dynamics for Hyperparameter Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
4.1 Algorithmic update for surrogate model
In this section, Hyperband is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Algorithmic update for surrogate model" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Hyperband is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Hyperband can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Hyperband affects optimization but the model remains interpretable.
- A transformer training diagnostic where Hyperband appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Hyperband as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Hyperband, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Hyperband visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Hyperband is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.2 Stability role of Gaussian process
In this section, ASHA is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Stability role of Gaussian process" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, ASHA is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where ASHA can be computed directly and compared with theory.
- A logistic-regression or softmax objective where ASHA affects optimization but the model remains interpretable.
- A transformer training diagnostic where ASHA appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating ASHA as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving ASHA, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes ASHA visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about ASHA is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.3 Rate or complexity controlled by expected improvement
In this section, BOHB is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Rate or complexity controlled by expected improvement" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, BOHB is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where BOHB can be computed directly and compared with theory.
- A logistic-regression or softmax objective where BOHB affects optimization but the model remains interpretable.
- A transformer training diagnostic where BOHB appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating BOHB as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving BOHB, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes BOHB visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about BOHB is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.4 Diagnostic interpretation of the update path
In this section, population-based training is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Diagnostic interpretation of the update path" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, population-based training is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where population-based training can be computed directly and compared with theory.
- A logistic-regression or softmax objective where population-based training affects optimization but the model remains interpretable.
- A transformer training diagnostic where population-based training appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating population-based training as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving population-based training, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes population-based training visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about population-based training is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.5 Connection to the next section in the chapter
In this section, multi-objective tuning is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Hyperparameter Optimization, the phrase "Connection to the next section in the chapter" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, multi-objective tuning is the part of Hyperparameter Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where multi-objective tuning can be computed directly and compared with theory.
- A logistic-regression or softmax objective where multi-objective tuning affects optimization but the model remains interpretable.
- A transformer training diagnostic where multi-objective tuning appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating multi-objective tuning as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving multi-objective tuning, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes multi-objective tuning visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about multi-objective tuning is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- learning-rate, weight-decay, batch-size, and schedule tuning for LLM fine-tuning.
- Hyperband and ASHA for neural architecture and training-budget search.
- Bayesian optimization for expensive, low-dimensional continuous tuning.
- validation leakage prevention in model-selection pipelines.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.