This handout is not an introduction. It is written for those who have already conquered the standard curriculum — you can differentiate in your sleep, convexity is second nature, and AM-GM is a reflex. If you’re here, you’ve cleared the floor that 99% never reach. What follows is the architecture above it: the theorems, algorithms, and structural insights that separate the top solvers from everyone who merely “knows the material.” Proceed accordingly.
In Part 1 , we laid the algebraic and topological foundations. In Part 2 , we forged the differential tools — recursive derivatives, the wavy curve, and sign analysis. In Part 3 , we deployed the inequality machinery — the Mean Value Theorems, convexity, and a six-technique heuristic arsenal. Now we ascend to the full landscape.
This is where the top 1% separate from the merely excellent. The tools in this module — Schur convexity, Muirhead’s theorem, the Sum of Squares algorithm, Vasc’s Equal Variables method, the L^p integral bounds (Hölder, Minkowski, Young), and the Rearrangement inequality — are the weapons that obliterate problems which resist every classical approach. Each framework here has destroyed entire generations of Olympiad contestants who lacked the vocabulary to even name the technique they needed.
I. The Majorization Deepening
In Part 3, Section II , we defined majorization and stated Karamata’s inequality. Here we excavate the structural foundations — the matrix-theoretic characterization that reveals why majorization works, and the differential criterion that makes Schur convexity computable.
The Hardy-Littlewood-Pólya Theorem
The vector majorizes (i.e., ) if and only if there exists a doubly stochastic matrix such that .
A doubly stochastic matrix is a square matrix of non-negative reals where every row and every column sums to exactly .
Why this matters: The theorem says that being majorized by is equivalent to being a “weighted average” of the components of . The vector is literally a smoothed version of .
Birkhoff’s Theorem and the Geometric Picture
The set of all doubly stochastic matrices forms a convex polytope whose extreme points (vertices) are precisely the permutation matrices.
Combining Hardy-Littlewood-Pólya with Birkhoff: is majorized by if and only if can be written as a convex combination of permutations of . In competition terms, if you can show one vector is a convex mix of permutations of another, majorization is immediate.
Schur Convexity
A symmetric function is Schur-convex if implies .
Equivalently, is Schur-convex if it preserves the majorization ordering.
A continuously differentiable symmetric function is Schur-convex if and only if for all pairs :
To prove where :
- Verify symmetry: Confirm is symmetric under variable permutation.
- Compute partials: Find and .
- Factor: Compute the difference and factor out .
- Sign check: If the remaining factor is globally non-negative, is Schur-convex.
The collapse: Once Schur-convexity is established, the minimum of under a fixed sum constraint occurs at the completely averaged vector , and the maximum at the maximally skewed vector .
II. Symmetric Polynomial Bounds
Muirhead’s Theorem
Let and be exponent sequences, and define the symmetric mean:
For all positive reals : if and only if .
| Exponent | Exponent | Majorization Check | Result |
|---|---|---|---|
| , , ✓ | |||
| , , ✓ | |||
| ✗ | Reversed — bound goes the other way |
In stringent grading environments (IMO, Putnam written), citing “by Muirhead” is strategically risky — it can be perceived as invoking a black box. Muirhead’s true competition value is as a heuristic oracle:
- Check feasibility: Does ? If not, the bound is impossible.
- Construct the proof: Build an explicit weighted AM-GM chain that evaluates to the same symmetric result.
- Submit the AM-GM proof: This is universally accepted without further justification.
The oracle tells you what is true. AM-GM proves why.
Muirhead’s theorem requires the full symmetric sum (averaging over all permutations). A cyclic sum over three variables like involves only 3 terms, not the terms of the symmetric sum .
Applying majorization logic to cyclic permutations produces mathematically false results. This error on an Olympiad paper results in immediate penalization.
The Sum of Squares (SOS) Method
For a symmetric inequality with three variables, the SOS decomposition writes:
Since , proving reduces to analyzing the coefficient signs .
Assuming WLOG , the algebraic identity (since ) yields the following positivity criteria:
| Known Signs | Required Criterion |
|---|---|
| None — trivially non-negative | |
| , , | |
| , , | |
| and | and |
- Expand and collect terms.
- Extract squared differences: use the identity and similar.
- Collect coefficients of , , across all cyclic shifts.
- Check the SOS positivity criteria under the ordering .
Power: SOS handles asymmetric fractional denominators and radical roots by systematically multiplying out conjugates to extract factors.
III. Calculus-Driven Optimization
Vasc’s Equal Variables (EV) Method
While SOS handles algebraic reductions, Vasile Cîrtoaje’s Equal Variables method provides a calculus-driven topological approach. It resolves optimization bounds by proving that extremal values of symmetric functions under sum or product constraints occur when at least two variables are equal.
Let have exactly one inflection point , with convex on and concave on . Then the extrema of subject to occur when at least of the variables are equal.
For a symmetric inequality with :
- Identify the component function: Write or relate to such a sum.
- Find the inflection point: Compute and locate the sign change.
- Apply the Half-Convex theorem: The minimum occurs at or for some values.
- Substitute: Set , use , and reduce to a single-variable polynomial.
- Factor: The resulting polynomial typically factors as with .
Why this is devastating: It systematically obliterates symmetric inequalities up to degree 8 that completely resist AM-GM, Schur, or Jensen.
When applying Karamata’s inequality, sequences must be sorted in descending order before checking partial sum conditions. If and , checking yields — false. Sort first to , , then and . ✓
IV. L^p Integral Bounds
The discrete algebraic frameworks above govern the finite sequences of Olympiad algebra. The continuous counterparts — governing integrals over measure spaces — dominate the subjective calculus sections of ISI B.Math, CMI entrance, and advanced Putnam problems. Three inequalities form the complete hierarchy.
Young’s Inequality
Let satisfy the conjugate relation . For any non-negative reals :
The geometric proof: Consider a strictly increasing function with inverse . The rectangle with vertices at the origin, , , has area . This area is bounded by the sum of the area under from to and the area to the left of from to :
This visceral geometric argument — bounding a rectangle by two complementary curved regions — is the deepest way to understand why products are bounded by scaled powers.
Hölder’s Inequality
For measurable functions and conjugate exponents with :
Special case: recovers the Cauchy-Schwarz inequality for integrals.
The standard proof of Hölder’s inequality uses integral normalization — a meta-technique that appears repeatedly in ISI/CMI subjective problems:
- Define normalized functions: and .
- This scaling forces and .
- Apply Young’s pointwise: .
- Integrate: .
- Unpack the normalization to recover the standard Hölder form.
When to deploy: Any integral problem where you need to bound by separate norms of and .
Minkowski’s Inequality
For and measurable functions :
This establishes that the L^p norm satisfies the triangle inequality, making a Banach space.
The derivation cascade: Factor , bound via triangle inequality, apply Hölder to each resulting integral, factor out , and divide. The exponent arithmetic closes the proof.
| Framework | Domain | Core Statement | Primary Use |
|---|---|---|---|
| Young’s | Pointwise | Converting products to additive bounds | |
| Hölder’s | Integral | Bounding integrated products (ISI subjective) | |
| Minkowski’s | Integral | Triangle inequality for functional interpolation |
V. Permutation Optimization
The Rearrangement Inequality
Given sorted sequences and , and any permutation :
The dot product is maximized when the sequences are similarly sorted and minimized when oppositely sorted.
The swapping proof: Assume a permutation achieves the maximum but is not the identity. Then there exist with . Since and , swapping the pairings changes the sum by . Repeating until no misaligned pairs remain forces the identity permutation.
Chebyshev’s Sum Inequality
If and are similarly sorted, then:
For oppositely sorted sequences, the inequality reverses.
Derivation from Rearrangement: Consider the cyclic shifts of . By Rearrangement, the identity permutation sum dominates each cyclic shift. Summing all inequalities yields .
Before applying Rearrangement or Chebyshev:
- Explicitly establish sorting: Prove and determine the sorting of .
- Watch for entanglement: If the rank ordering of depends on a parameter that simultaneously alters the ordering of , the inequality cannot be directly invoked.
- Construct auxiliary sequences: When entanglement exists, define to artificially force monotonic alignment before applying the inequality.
Grading requirement: In written Olympiads, establishing the explicit ordering proof is mandatory. Skipping it produces a “fake-solve.”
Error: Applying the Rearrangement inequality to sequences whose monotonic ordering depends on an algebraic parameter that simultaneously affects both sequences.
If the sorting of changes with a parameter that also changes the sorting of , the inequality’s hypothesis is violated. Advanced Putnam problems deliberately exploit this trap.
Fix: Construct auxiliary sequences with provably fixed orderings before invoking the inequality.
VI. Advanced Manipulation Heuristics
Three algebraic micro-heuristics allow the major theorems above to bypass strict structural constraints.
Building on the tangent line method from Part 3, Section III , isolated fudging extends the technique to problems where global convexity fails:
- Identify the equality case: For a constrained sum , equality typically occurs at for all .
- Compute the tangent line: .
- Prove the algebraic bound directly — by factoring the difference polynomial. Often divides the difference.
- Sum: since .
The key difference from Part 3: This works even when changes sign — you bypass convexity entirely by proving the tangent bound algebraically via polynomial factorization.
Inequalities conditioned on a constant product (e.g., ) often prevent polynomial degree matching. Homogenize via:
This ensures unconditionally for all , freeing the inequality for attack by Muirhead’s or Schur’s theorems without the product constraint interfering.
When to use: Any time a product constraint prevents direct degree comparison or AM-GM alignment.
For structural conditions like , the substitution , , — where are angles of a triangle — ensures the constraint is inherently satisfied.
This transforms rigid algebraic bounds into trigonometric optimization, opening the door for Jensen’s inequality on concave functions like and .
Signal: When you see appearing as a constraint alongside expressions involving , the tangent half-angle substitution is almost certainly the intended path.
Error: Attempting to apply Muirhead’s theorem to an inequality whose left and right sides have different polynomial degrees.
Muirhead compares symmetric sums of the same total degree. If the LHS has degree 7 and the RHS has degree 5, you must homogenize by multiplying the lower-degree side by (using the constraint) to equalize degrees before applying the theorem.
Fix: Always verify degree matching before invoking Muirhead. Use the constraint to embed the necessary powers.
Error: Writing “by Muirhead’s theorem” as the sole justification in a written Olympiad proof.
In stringent grading environments (IMO, Putnam), this can be penalized as invoking an unproven result. Muirhead’s theorem is not on the “universally accepted without proof” list (unlike AM-GM).
Fix: Use Muirhead as a heuristic oracle to identify the correct bound, then construct an explicit AM-GM chain that proves it.
VII. Beyond the Syllabus
The theory of majorization extends far beyond real vectors. For Hermitian matrices and , the vector of eigenvalues of is majorized by the sum of eigenvalue vectors of and . This is proved via the Rayleigh quotient characterization of eigenvalues and Birkhoff’s polytope machinery. Lieb’s Concavity Theorem — that certain trace-exponential functions of Hermitian matrices preserve concavity — is the frontier, connecting the convexity hierarchy of Part 3 to quantum information theory.
The symmetric means from Muirhead’s theorem are special cases of Schur polynomials — the characters of irreducible representations of the general linear group . The majorization ordering on exponent vectors corresponds precisely to the dominance ordering on partitions, which governs the decomposition of tensor products of representations. When a competition problem asks you to compare symmetric polynomial sums, you are — perhaps unknowingly — navigating the lattice of Young diagrams.
In probability theory, the L^p bounds (Hölder, Minkowski) are the backbone of concentration inequalities — results that show random variables cluster near their expectations. Markov’s inequality () is a direct consequence of Hölder with , . The Chernoff bound, which governs tail probabilities of independent sums, uses the exponential moment technique: apply Hölder to . These tools are the bridge between the calculus inequalities of this module and the probabilistic world of information theory, statistical mechanics, and machine learning.
Illustrative Examples
(Isolated Fudging / Tangent Line Trick)
Let with . Prove that .
View Solution
Step 1 — Identify the equality case: By symmetry and the constraint, equality should occur at .
Step 2 — Define the component function: Let . The inequality becomes .
Step 3 — Compute the tangent at the equality point: . , so .
The tangent line at : .
Step 4 — Prove algebraically: The difference is:
Since the tangent touches at , the factor must divide the numerator. By polynomial division:
For : and , so the product is non-negative. ✓
Step 5 — Sum over the variables:
Equality if and only if .
(The SOS Algorithm)
For positive reals , prove that .
View Solution
Step 1 — Extract SOS structure: Use the identity .
Substituting into the cyclic sum and collecting coefficients of across all three permutations yields the canonical SOS form .
Step 2 — Compute : Tracking the coefficient across cyclic shifts:
Step 3 — Sign analysis under : This ordering implies:
Denominators shrink, so reciprocals grow. Therefore , guaranteeing . By identical reasoning, .
Step 4 — Handle the potentially negative : The SOS criterion requires . Computing:
After cancellation: since .
Step 5 — Conclude: All SOS positivity criteria are satisfied. Therefore . Equality if and only if .
(Vasc’s Equal Variables Method)
Let with . Prove that .
View Solution
Step 1 — Setup: Define . We need .
Step 2 — Apply the mixing variables displacement: WLOG . Let and , so and . Compute .
Using the algebraic expansions:
Step 3 — Compute the displacement:
Step 4 — Reduce to equal variables: By the Half-Convex Function Theorem, it suffices to prove where , i.e., . Substituting:
Step 5 — Factor the single-variable polynomial: Expanding and simplifying:
Since always, and has no real roots, the product is strictly non-negative.
Step 6 — Conclude: The global minimum occurs at , where . Therefore unconditionally.
(Muirhead’s Theorem with Homogenization)
For positive reals with , prove that .
Direct Muirhead on the cyclic sum
The LHS is a cyclic sum, not a symmetric sum. Muirhead’s theorem requires averaging over all permutations, not just the 3 cyclic ones. Attempting to apply Muirhead directly here is the Cyclic Sum Trap (Trap 1) — the result would be mathematically invalid.
View Solution
Step 1 — Eliminate denominators: Since , we have , so . Therefore:
The inequality transforms to .
Step 2 — Homogenize to equal degree: The LHS has degree . The RHS has degree 2. Multiply the RHS by to match: .
Step 3 — Symmetrize the LHS: The symmetric sum has exponent vector and the target has .
Step 4 — Verify majorization:
- ✓
- ✓
- ✓
Since , Muirhead’s theorem gives . (Note: Muirhead’s theorem extends to real — not just integer — exponents for positive variables, as the symmetric mean is well-defined for all when .)
Step 5 — Construct the AM-GM chain: For the constructive proof, by weighted AM-GM:
Summing cyclically over all three variables completes the proof.
(Continuous Inequality via Monotonicity and Integration)
Let be differentiable with and . Prove that for all .
View Solution
Step 1 — Establish monotonicity: Since on and , the denominator . Thus — the function is strictly increasing.
Step 2 — Bound from below: Since is increasing and , we have for all . Therefore .
Step 3 — Bound from above: Adding to both sides of :
Step 4 — Integrate: For any :
Step 5 — Bound the arctangent: Since is bounded above by :
Therefore for all .
(AM-GM on Geometric Series Terms)
For and positive integer , prove that .
View Solution
Step 1 — Recognize the sum structure: The LHS is the geometric series sum:
Step 2 — Apply AM-GM: The arithmetic mean of the terms is bounded below by their geometric mean:
Step 3 — Compute the geometric mean: The exponent sum is . Thus:
Step 4 — Conclude: Multiplying both sides by :
Equality holds if and only if all terms are equal, i.e., (verified via L’Hôpital: ).
(Rearrangement Inequality with Trigonometric Substitution)
Let with . Prove that .
View Solution
Step 1 — Decode the constraint: Since , for each variable implies (as for positive ). More precisely, multiplying the original inequality by gives , but the key structural observation is that by AM-GM on each pair , all variables satisfy . The tangent half-angle substitution below handles the constraint algebraically: the condition for positive reals maps to .
Step 2 — Trigonometric substitution: Let , , where . The constraint corresponds to — the angles form an acute triangle.
Step 3 — Simplify the target expression: For each term:
The inequality becomes .
Step 4 — Apply Jensen’s Inequality: Define . Then on , so is strictly concave.
By Jensen for concave functions:
Since and is increasing on :
(Deriving Minkowski’s Inequality from Hölder’s)
Using Hölder’s inequality as the sole base bounding mechanism, rigorously derive Minkowski’s inequality: for .
View Solution
Step 1 — Factor the exponent: Write . By the triangle inequality, , so:
Step 2 — Apply Hölder to each term: Let be the Hölder conjugate of , i.e., , which gives . Applying Hölder’s inequality to the first integral:
By identical logic for the term: .
Step 3 — Combine and simplify: Adding both bounds:
Step 4 — Divide by the common factor: Dividing both sides by (valid when non-zero):
Since , the left side reduces to . Therefore:
Selected Problems
Let be positive reals with . Prove that for any .
Hint
For positive reals , prove that .
Hint
Let be a polynomial with complex roots . If , find the minimum possible average distance of the roots to the origin, i.e., minimize .
Hint
Let with . Prove that .
Hint
Prove Karamata’s inequality for explicitly: if is convex and , then .
Hint
For non-negative reals , prove that .
Hint
Suppose are positive definite matrices. Prove that the vector of eigenvalues of is majorized by the sum of the eigenvalue vectors of and .
Hint
Let be continuous and non-negative on . Find the sharp upper bound for .
Hint
Challenge Problem
(The Synthesis Problem)
Let be strictly positive real numbers. Prove that: