In convex analysis and the calculus of variations, both branches of mathematics, a pseudoconvex function is a function that behaves like a convex function with respect to finding its local minima, but need not actually be convex. Informally, a differentiable function is pseudoconvex if it is increasing in any direction where it has a positive directional derivative. The property must hold in all of the function domain, and not only for nearby points.
Consider a differentiable function f : X ⊆ R n → R {\displaystyle f:X\subseteq \mathbb {R} ^{n}\rightarrow \mathbb {R} } , defined on a (nonempty) convex open set X {\displaystyle X} of the finite-dimensional Euclidean space R n {\displaystyle \mathbb {R} ^{n}} . This function is said to be pseudoconvex if the following property holds:[1]
Equivalently:
Here ∇ f {\displaystyle \nabla f} is the gradient of f {\displaystyle f} , defined by: ∇ f = ( ∂ f ∂ x 1 , … , ∂ f ∂ x n ) . {\displaystyle \nabla f=\left({\frac {\partial f}{\partial x_{1}}},\dots ,{\frac {\partial f}{\partial x_{n}}}\right).}
Note that the definition may also be stated in terms of the directional derivative of f {\displaystyle f} , in the direction given by the vector v = y − x {\displaystyle v=y-x} . This is because, as f {\displaystyle f} is differentiable, this directional derivative is given by:
Every convex function is pseudoconvex, but the converse is not true. For example, the function f ( x ) = x + x 3 {\displaystyle f(x)=x+x^{3}} is pseudoconvex but not convex. Similarly, any pseudoconvex function is quasiconvex; but the converse is not true, since the function f ( x ) = x 3 {\displaystyle f(x)=x^{3}} is quasiconvex but not pseudoconvex. This can be summarized schematically as:
To see that f ( x ) = x 3 {\displaystyle f(x)=x^{3}} is not pseudoconvex, consider its derivative at x = 0 {\displaystyle x=0} : f ′ ( 0 ) = 0 {\displaystyle f^{\prime }(0)=0} . Then, if f ( x ) = x 3 {\displaystyle f(x)=x^{3}} was pseudoconvex, we should have:
In particular it should be true for y = − 1 {\displaystyle y=-1} . But it is not, as: f ( − 1 ) = ( − 1 ) 3 = − 1 < f ( 0 ) = 0 {\displaystyle f(-1)=(-1)^{3}=-1<f(0)=0} .
For any differentiable function, we have the Fermat's theorem necessary condition of optimality, which states that: if f {\displaystyle f} has a local minimum at x ∗ {\displaystyle x^{*}} in an open domain, then x ∗ {\displaystyle x^{*}} must be a stationary point of f {\displaystyle f} (that is: ∇ f ( x ∗ ) = 0 {\displaystyle \nabla f(x^{*})=0} ).
Pseudoconvexity is of great interest in the area of optimization, because the converse is also true for any pseudoconvex function. That is:[2] if x ∗ {\displaystyle x^{*}} is a stationary point of a pseudoconvex function f {\displaystyle f} , then f {\displaystyle f} has a global minimum at x ∗ {\displaystyle x^{*}} . Note also that the result guarantees a global minimum (not only local).
This last result is also true for a convex function, but it is not true for a quasiconvex function. Consider for example the quasiconvex function:
This function is not pseudoconvex, but it is quasiconvex. Also, the point x = 0 {\displaystyle x=0} is a critical point of f {\displaystyle f} , as f ′ ( 0 ) = 0 {\displaystyle f^{\prime }(0)=0} . However, f {\displaystyle f} does not have a global minimum at x = 0 {\displaystyle x=0} (not even a local minimum).
Finally, note that a pseudoconvex function may not have any critical point. Take for example the pseudoconvex function: f ( x ) = x 3 + x {\displaystyle f(x)=x^{3}+x} , whose derivative is always positive: f ′ ( x ) = 3 x 2 + 1 > 0 , ∀ x ∈ R {\displaystyle f^{\prime }(x)=3x^{2}+1>0,\,\forall \,x\in \mathbb {R} } .
An example of a function that is pseudoconvex, but not convex, is: f ( x ) = x 2 x 2 + k , k > 0. {\displaystyle f(x)={\frac {x^{2}}{x^{2}+k}},\,k>0.} The figure shows this function for the case where k = 0.2 {\displaystyle k=0.2} . This example may be generalized to two variables as:
The previous example may be modified to obtain a function that is not convex, nor pseudoconvex, but is quasiconvex:
The figure shows this function for the case where k = 0.5 , p = 0.6 {\displaystyle k=0.5,p=0.6} . As can be seen, this function is not convex because of the concavity, and it is not pseudoconvex because it is not differentiable at x = 0 {\displaystyle x=0} .
The notion of pseudoconvexity can be generalized to nondifferentiable functions as follows.[3] Given any function f : X → R {\displaystyle f:X\rightarrow \mathbb {R} } , we can define the upper Dini derivative of f {\displaystyle f} by:
where u is any unit vector. The function is said to be pseudoconvex if it is increasing in any direction where the upper Dini derivative is positive. More precisely, this is characterized in terms of the subdifferential ∂ f {\displaystyle \partial f} as follows:
where [ x , y ] {\displaystyle [x,y]} denotes the line segment adjoining x and y.
A pseudoconcave function is a function whose negative is pseudoconvex. A pseudolinear function is a function that is both pseudoconvex and pseudoconcave.[4] For example, linear–fractional programs have pseudolinear objective functions and linear–inequality constraints. These properties allow fractional-linear problems to be solved by a variant of the simplex algorithm (of George B. Dantzig).[5][6][7]
Given a vector-valued function η {\displaystyle \eta } , there is a more general notion of η {\displaystyle \eta } -pseudoconvexity[8][9] and η {\displaystyle \eta } -pseudolinearity; wherein classical pseudoconvexity and pseudolinearity pertain to the case when η ( x , y ) = y − x {\displaystyle \eta (x,y)=y-x} .