
- Home
- Introduction
- Linear Programming
- Norm
- Inner Product
- Minima and Maxima
- Convex Set
- Affine Set
- Convex Hull
- Caratheodory Theorem
- Weierstrass Theorem
- Closest Point Theorem
- Fundamental Separation Theorem
- Convex Cones
- Polar Cone
- Conic Combination
- Polyhedral Set
- Extreme point of a convex set
- Direction
- Convex & Concave Function
- Jensen's Inequality
- Differentiable Convex Function
- Sufficient & Necessary Conditions for Global Optima
- Quasiconvex & Quasiconcave functions
- Differentiable Quasiconvex Function
- Strictly Quasiconvex Function
- Strongly Quasiconvex Function
- Pseudoconvex Function
- Convex Programming Problem
- Fritz-John Conditions
- Karush-Kuhn-Tucker Optimality Necessary Conditions
- Algorithms for Convex Problems
Convex Optimization - Differentiable Function
Let S be a non-empty open set in Rn,then f:S→R is said to be differentiable at ˆx∈S if there exist a vector ▽f(ˆx) called gradient vector and a function α:Rn→R such that
f(x)=f(ˆx)+▽f(ˆx)T(x−ˆx)+‖x=ˆx‖α(ˆx,x−ˆx),∀x∈S where
α(ˆx,x−ˆx)→0▽f(ˆx)=[∂f∂x1∂f∂x2...∂f∂xn]Tx=ˆx
Theorem
let S be a non-empty, open convexset in Rn and let f:S→R be differentiable on S. Then, f is convex if and only if for x1,x2∈S,▽f(x2)T(x1−x2)≤f(x1)−f(x2)
Proof
Let f be a convex function. i.e., for x1,x2∈S,λ∈(0,1)
f[λx1+(1−λ)x2]≤λf(x1)+(1−λ)f(x2)
⇒f[λx1+(1−λ)x2]≤λ(f(x1)−f(x2))+f(x2)
⇒λ(f(x1)−f(x2))≥f(x2+λ(x1−x2))−f(x2)
⇒λ(f(x1)−f(x2))≥f(x2)+▽f(x2)T(x1−x2)λ+
‖λ(x1−x2)‖α(x2,λ(x1−x2)−f(x2))
where α(x2,λ(x1−x2))→0 asλ→0
Dividing by λ on both sides, we get −
f(x1)−f(x2)≥▽f(x2)T(x1−x2)
Converse
Let for x1,x2∈S,▽f(x2)T(x1−x2)≤f(x1)−f(x2)
To show that f is convex.
Since S is convex, x3=λx1+(1−λ)x2∈S,λ∈(0,1)
Since x1,x3∈S, therefore
f(x1)−f(x3)≥▽f(x3)T(x1−x3)
⇒f(x1)−f(x3)≥▽f(x3)T(x1−λx1−(1−λ)x2)
⇒f(x1)−f(x3)≥(1−λ)▽f(x3)T(x1−x2)
Since, x2,x3∈S therefore
f(x2)−f(x3)≥▽f(x3)T(x2−x3)
⇒f(x2)−f(x3)≥▽f(x3)T(x2−λx1−(1−λ)x2)
⇒f(x2)−f(x3)≥(−λ)▽f(x3)T(x1−x2)
Thus, combining the above equations, we get −
λ(f(x1)−f(x3))+(1−λ)(f(x2)−f(x3))≥0
⇒f(x3)≤λf(x1)+(1−λ)f(x2)
Theorem
let S be a non-empty open convex set in Rn and let f:S→R be differentiable on S, then f is convex on S if and only if for any x1,x2∈S,(▽f(x2)−▽f(x1))T(x2−x1)≥0
Proof
let f be a convex function, then using the previous theorem −
▽f(x2)T(x1−x2)≤f(x1)−f(x2) and
▽f(x1)T(x2−x1)≤f(x2)−f(x1)
Adding the above two equations, we get −
▽f(x2)T(x1−x2)+▽f(x1)T(x2−x1)≤0
⇒(▽f(x2)−▽f(x1))T(x1−x2)≤0
⇒(▽f(x2)−▽f(x1))T(x2−x1)≥0
Converse
Let for any x1,x2∈S,(▽f(x2)−▽f(x1))T(x2−x1)≥0
To show that f is convex.
Let x1,x2∈S, thus by mean value theorem, f(x1)−f(x2)x1−x2=▽f(x),x∈(x1−x2)⇒x=λx1+(1−λ)x2 because S is a convex set.
⇒f(x1)−f(x2)=(▽f(x)T)(x1−x2)
for x,x1, we know −
(▽f(x)−▽f(x1))T(x−x1)≥0
⇒(▽f(x)−▽f(x1))T(λx1+(1−λ)x2−x1)≥0
⇒(▽f(x)−▽f(x1))T(1−λ)(x2−x1)≥0
⇒▽f(x)T(x2−x1)≥▽f(x1)T(x2−x1)
Combining the above equations, we get −
⇒▽f(x1)T(x2−x1)≤f(x2)−f(x1)
Hence using the last theorem, f is a convex function.
Twice Differentiable function
Let S be a non-empty subset of Rn and let f:S→R then f is said to be twice differentiable at ˉx∈S if there exists a vector ▽f(ˉx),anXn matrix H(x)calledHessianmatrix and a function α:Rn→R such that f(x)=f(ˉx+x−ˉx)=f(ˉx)+▽f(ˉx)T(x−ˉx)+12(x−ˉx)H(ˉx)(x−ˉx)
where α(ˉx,x−ˉx)→Oasx→ˉx