Lesson overview | Previous part | Next part
Constrained Optimization: Part 3: Core Theory I: Geometry and Guarantees to 4. Core Theory II: Algorithms and Dynamics
3. Core Theory I: Geometry and Guarantees
This block develops core theory i: geometry and guarantees for Constrained Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
3.1 Geometry of inequality constraints
In this section, constraint qualifications is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Geometry of inequality constraints" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, constraint qualifications is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where constraint qualifications can be computed directly and compared with theory.
- A logistic-regression or softmax objective where constraint qualifications affects optimization but the model remains interpretable.
- A transformer training diagnostic where constraint qualifications appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating constraint qualifications as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving constraint qualifications, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes constraint qualifications visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about constraint qualifications is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.2 Key inequality for Lagrangian
In this section, Slater condition is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Key inequality for Lagrangian" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Slater condition is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Slater condition can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Slater condition affects optimization but the model remains interpretable.
- A transformer training diagnostic where Slater condition appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Slater condition as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Slater condition, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Slater condition visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Slater condition is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.3 Role of stationarity
In this section, dual problem is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Role of stationarity" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, dual problem is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where dual problem can be computed directly and compared with theory.
- A logistic-regression or softmax objective where dual problem affects optimization but the model remains interpretable.
- A transformer training diagnostic where dual problem appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating dual problem as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving dual problem, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes dual problem visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about dual problem is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.4 Proof template and what the proof actually buys
In this section, projected gradient descent is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Proof template and what the proof actually buys" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, projected gradient descent is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where projected gradient descent can be computed directly and compared with theory.
- A logistic-regression or softmax objective where projected gradient descent affects optimization but the model remains interpretable.
- A transformer training diagnostic where projected gradient descent appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating projected gradient descent as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving projected gradient descent, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes projected gradient descent visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about projected gradient descent is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
3.5 Failure modes when assumptions are removed
In this section, Euclidean projection is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Failure modes when assumptions are removed" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Euclidean projection is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Euclidean projection can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Euclidean projection affects optimization but the model remains interpretable.
- A transformer training diagnostic where Euclidean projection appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Euclidean projection as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Euclidean projection, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Euclidean projection visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Euclidean projection is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4. Core Theory II: Algorithms and Dynamics
This block develops core theory ii: algorithms and dynamics for Constrained Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
4.1 Algorithmic update for primal feasibility
In this section, projected gradient descent is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Algorithmic update for primal feasibility" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, projected gradient descent is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where projected gradient descent can be computed directly and compared with theory.
- A logistic-regression or softmax objective where projected gradient descent affects optimization but the model remains interpretable.
- A transformer training diagnostic where projected gradient descent appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating projected gradient descent as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving projected gradient descent, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes projected gradient descent visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about projected gradient descent is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.2 Stability role of dual feasibility
In this section, Euclidean projection is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Stability role of dual feasibility" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Euclidean projection is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Euclidean projection can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Euclidean projection affects optimization but the model remains interpretable.
- A transformer training diagnostic where Euclidean projection appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Euclidean projection as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Euclidean projection, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Euclidean projection visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Euclidean projection is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.3 Rate or complexity controlled by complementary slackness
In this section, penalty methods is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Rate or complexity controlled by complementary slackness" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, penalty methods is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where penalty methods can be computed directly and compared with theory.
- A logistic-regression or softmax objective where penalty methods affects optimization but the model remains interpretable.
- A transformer training diagnostic where penalty methods appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating penalty methods as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving penalty methods, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes penalty methods visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about penalty methods is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.4 Diagnostic interpretation of the update path
In this section, barrier methods is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Diagnostic interpretation of the update path" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, barrier methods is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where barrier methods can be computed directly and compared with theory.
- A logistic-regression or softmax objective where barrier methods affects optimization but the model remains interpretable.
- A transformer training diagnostic where barrier methods appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating barrier methods as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving barrier methods, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes barrier methods visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about barrier methods is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
4.5 Connection to the next section in the chapter
In this section, augmented Lagrangian is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Constrained Optimization, the phrase "Connection to the next section in the chapter" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, augmented Lagrangian is the part of Constrained Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where augmented Lagrangian can be computed directly and compared with theory.
- A logistic-regression or softmax objective where augmented Lagrangian affects optimization but the model remains interpretable.
- A transformer training diagnostic where augmented Lagrangian appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating augmented Lagrangian as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving augmented Lagrangian, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes augmented Lagrangian visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about augmented Lagrangian is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- support-vector machines through KKT and dual variables.
- fairness, safety, and resource-constrained model training.
- projection layers for nonnegative or norm-constrained parameters.
- ADMM-style splitting for distributed and federated objectives.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.