Lesson overview | Previous part | Next part
Convex Optimization: Part 5: Core Theory III: Practical Variants to 6. Advanced Topics
5. Core Theory III: Practical Variants
This block develops core theory iii: practical variants for Convex Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
5.1 Variant built around convex problem classes
In this section, weak duality is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Variant built around convex problem classes" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, weak duality is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where weak duality can be computed directly and compared with theory.
- A logistic-regression or softmax objective where weak duality affects optimization but the model remains interpretable.
- A transformer training diagnostic where weak duality appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating weak duality as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving weak duality, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes weak duality visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about weak duality is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.2 Variant built around linear programs
In this section, strong duality is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Variant built around linear programs" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, strong duality is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where strong duality can be computed directly and compared with theory.
- A logistic-regression or softmax objective where strong duality affects optimization but the model remains interpretable.
- A transformer training diagnostic where strong duality appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating strong duality as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving strong duality, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes strong duality visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about strong duality is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.3 Variant built around quadratic programs
In this section, Slater condition is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Variant built around quadratic programs" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, Slater condition is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where Slater condition can be computed directly and compared with theory.
- A logistic-regression or softmax objective where Slater condition affects optimization but the model remains interpretable.
- A transformer training diagnostic where Slater condition appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating Slater condition as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving Slater condition, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes Slater condition visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about Slater condition is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.4 Implementation constraints and numerical stability
In this section, optimality certificates is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Implementation constraints and numerical stability" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, optimality certificates is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where optimality certificates can be computed directly and compared with theory.
- A logistic-regression or softmax objective where optimality certificates affects optimization but the model remains interpretable.
- A transformer training diagnostic where optimality certificates appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating optimality certificates as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving optimality certificates, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes optimality certificates visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about optimality certificates is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
5.5 What belongs here versus neighboring sections
In this section, logistic regression convexity is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "What belongs here versus neighboring sections" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, logistic regression convexity is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where logistic regression convexity can be computed directly and compared with theory.
- A logistic-regression or softmax objective where logistic regression convexity affects optimization but the model remains interpretable.
- A transformer training diagnostic where logistic regression convexity appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating logistic regression convexity as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving logistic regression convexity, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes logistic regression convexity visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about logistic regression convexity is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6. Advanced Topics
This block develops advanced topics for Convex Optimization. It keeps the scope local to this section while pointing forward when a neighboring topic owns the full treatment.
6.1 Advanced view of semidefinite programs
In this section, optimality certificates is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Advanced view of semidefinite programs" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, optimality certificates is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where optimality certificates can be computed directly and compared with theory.
- A logistic-regression or softmax objective where optimality certificates affects optimization but the model remains interpretable.
- A transformer training diagnostic where optimality certificates appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating optimality certificates as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving optimality certificates, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes optimality certificates visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about optimality certificates is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.2 Advanced view of subgradients
In this section, logistic regression convexity is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Advanced view of subgradients" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, logistic regression convexity is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where logistic regression convexity can be computed directly and compared with theory.
- A logistic-regression or softmax objective where logistic regression convexity affects optimization but the model remains interpretable.
- A transformer training diagnostic where logistic regression convexity appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating logistic regression convexity as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving logistic regression convexity, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes logistic regression convexity visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about logistic regression convexity is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.3 Advanced view of proximal operators
In this section, nuclear norm relaxation is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Advanced view of proximal operators" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, nuclear norm relaxation is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where nuclear norm relaxation can be computed directly and compared with theory.
- A logistic-regression or softmax objective where nuclear norm relaxation affects optimization but the model remains interpretable.
- A transformer training diagnostic where nuclear norm relaxation appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating nuclear norm relaxation as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving nuclear norm relaxation, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes nuclear norm relaxation visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about nuclear norm relaxation is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.4 Infinite-dimensional or large-scale interpretation
In this section, convex sets is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Infinite-dimensional or large-scale interpretation" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, convex sets is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where convex sets can be computed directly and compared with theory.
- A logistic-regression or softmax objective where convex sets affects optimization but the model remains interpretable.
- A transformer training diagnostic where convex sets appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating convex sets as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving convex sets, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes convex sets visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about convex sets is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.
6.5 Open questions for frontier model training
In this section, convex combinations is treated as a concrete optimization object rather than a slogan. The goal is to understand how it changes the objective, the update rule, the convergence story, and the diagnostics a practitioner should inspect when training a modern model. For Convex Optimization, the phrase "Open questions for frontier model training" means a precise mathematical habit: state the assumptions, write the update, identify what can be measured, and connect the result to a real AI training decision.
Definition.
For this section, convex combinations is the part of Convex Optimization that controls how the objective, feasible region, or update rule behaves under the assumptions currently in force.
Symbolically, we track it through , , , , and any auxiliary state used by the algorithm.
Examples:
- A small synthetic quadratic where convex combinations can be computed directly and compared with theory.
- A logistic-regression or softmax objective where convex combinations affects optimization but the model remains interpretable.
- A transformer training diagnostic where convex combinations appears through gradient norms, update norms, curvature, or validation loss.
Non-examples:
- Treating convex combinations as a hyperparameter recipe without checking the objective assumptions.
- Inferring global behavior from one noisy minibatch when the section requires a population or full-batch statement.
Useful formula:
Proof sketch or reasoning pattern:
Start with the local model around , isolate the term involving convex combinations, and use the section assumptions to bound the change in objective value. If the assumption is geometric, the proof turns a picture into an inequality. If the assumption is stochastic, the proof takes conditional expectation before applying the bound. If the assumption is algorithmic, the proof checks that the proposed update is a descent, projection, or preconditioning step. This pattern is reusable across optimization theory.
Implementation consequence:
- Log a metric that makes convex combinations visible; otherwise a training run can fail while the scalar loss hides the cause.
- Compare the measured update with the mathematical update below before blaming data or architecture.
- Keep units straight: parameter norm, gradient norm, update norm, objective value, and validation metric are different objects.
Diagnostic questions:
- Which assumption about convex combinations is most fragile in the current training setup?
- What number would you log to catch the failure one thousand steps before divergence?
AI connection:
- logistic regression and softmax regression as convex baselines.
- support-vector machines through primal and dual convex programs.
- nuclear-norm relaxations behind low-rank matrix recovery and LoRA intuition.
- regularized empirical risk minimization with explicit certificates.
Local scope boundary: This subsection may reference neighboring material, but the full canonical treatment stays in its own folder. For example, stochastic gradient noise belongs to Stochastic Optimization, external schedule shapes belong to Learning Rate Schedules, and cross-entropy as an information measure belongs to Cross-Entropy.