The role of convexity on saddlepoint dynamics: Lyapunov function and robustness
Abstract
This paper studies the projected saddlepoint dynamics associated to a convexconcave function, which we term saddle function. The dynamics consists of gradient descent of the saddle function in variables corresponding to convexity and (projected) gradient ascent in variables corresponding to concavity. We examine the role that the local and/or global nature of the convexityconcavity properties of the saddle function plays in guaranteeing convergence and robustness of the dynamics. Under the assumption that the saddle function is twice continuously differentiable, we provide a novel characterization of the omegalimit set of the trajectories of this dynamics in terms of the diagonal blocks of the Hessian. Using this characterization, we establish global asymptotic convergence of the dynamics under local strong convexityconcavity of the saddle function. When strong convexityconcavity holds globally, we establish three results. First, we identify a Lyapunov function (that decreases strictly along the trajectory) for the projected saddlepoint dynamics when the saddle function corresponds to the Lagrangian of a general constrained convex optimization problem. Second, for the particular case when the saddle function is the Lagrangian of an equalityconstrained optimization problem, we show inputtostate stability of the saddlepoint dynamics by providing an ISS Lyapunov function. Third, we use the latter result to design an opportunistic statetriggered implementation of the dynamics. Various examples illustrate our results.
1 Introduction
Saddlepoint dynamics and its variations have been used extensively in the design and analysis of distributed feedback controllers and optimization algorithms in several domains, including power networks, network flow problems, and zerosum games. The analysis of the global convergence of this class of dynamics typically relies on some global strong/strict convexityconcavity property of the saddle function defining the dynamics. The main aim of this paper is to refine this analysis by unveiling two ways in which convexityconcavity of the saddle function plays a role. First, we show that local strong convexityconcavity is enough to conclude global asymptotic convergence, thus generalizing previous results that rely on global strong/strict convexityconcavity instead. Second, we show that, if global strong convexityconcavity holds, then one can identify a novel Lyapunov function for the projected saddlepoint dynamics for the case when the saddle function is the Lagrangian of a constrained optimization problem. This, in turn, implies a stronger form of convergence, that is, inputtostate stability (ISS) and has important implications in the practical implementation of the saddlepoint dynamics.
Literature review
The analysis of the convergence properties of (projected) saddlepoint dynamics to the set of saddle points goes back to [2, 3], motivated by the study of nonlinear programming and optimization. These works employed direct methods, examining the approximate evolution of the distance of the trajectories to the saddle point and concluding attractivity by showing it to be decreasing. Subsequently, motivated by the extensive use of the saddlepoint dynamics in congestion control problems, the literature on communication networks developed a Lyapunovbased and passivitybased asymptotic stability analysis, see e.g. [4] and references therein. Motivated by network optimization, more recent works [5, 6] have employed indirect, LaSalletype arguments to analyze asymptotic convergence. For this class of problems, the aggregate nature of the objective function and the local computability of the constraints make the saddlepoint dynamics corresponding to the Lagrangian naturally distributed. Many other works exploit this dynamics to solve network optimization problems for various applications, e.g., distributed convex optimization [6, 7], distributed linear programming [8], bargaining problems [9], and power networks [10, 11, 12, 13, 14]. Another area of application is game theory, where saddlepoint dynamics is applied to find the Nash equilibria of twoperson zerosum games [15, 16]. In the context of distributed optimization, the recent work [17] employs a (strict) Lyapunov function approach to ensure asymptotic convergence of saddlepointlike dynamics. The work [18] examines the asymptotic behavior of the saddlepoint dynamics when the set of saddle points is not asymptotically stable and, instead, trajectories exhibit oscillatory behavior. Our previous work has established global asymptotic convergence of the saddlepoint dynamics [19] and the projected saddlepoint dynamics [20] under global strict convexityconcavity assumptions. The works mentioned above require similar or stronger global assumptions on the convexityconcavity properties of the saddle function to ensure convergence. Our results here directly generalize the convergence properties reported above. Specifically, we show that traditional assumptions on the problem setup can be relaxed if convergence of the dynamics is the desired property: global convergence of the projected saddlepoint dynamics can be guaranteed under local strong convexityconcavity assumptions. Furthermore, if traditional assumptions do hold, then a stronger notion of convergence, that also implies robustness, is guaranteed: if strong convexityconcavity holds globally, the dynamics admits a Lyapunov function and in the absence of projection, the dynamics is ISS, admitting an ISS Lyapunov function.
Statement of contributions
Our starting point is the definition of the projected saddlepoint dynamics for a differentiable convexconcave function, referred to as saddle function. The dynamics has three components: gradient descent, projected gradient ascent, and gradient ascent of the saddle function, where each gradient is with respect to a subset of the arguments of the function. This unified formulation encompasses all forms of the saddlepoint dynamics mentioned in the literature review above. Our contributions shed light on the effect that the convexityconcavity of the saddle function has on the convergence attributes of the projected saddlepoint dynamics. Our first contribution is a novel characterization of the omegalimit set of the trajectories of the projected saddlepoint dynamics in terms of the diagonal Hessian blocks of the saddle function. To this end, we use the distance to a saddle point as a LaSalle function, express the Lie derivative of this function in terms of the Hessian blocks, and show it is nonpositive using secondorder properties of the saddle function. Building on this characterization, our second contribution establishes global asymptotic convergence of the projected saddlepoint dynamics to a saddle point assuming only local strong convexityconcavity of the saddle function. Our third contribution identifies a novel Lyapunov function for the projected saddlepoint dynamics for the case when strong convexityconcavity holds globally and the saddle function can be written as the Lagrangian of a constrained optimization problem. This discontinuous Lyapunov function can be interpreted as multiple continuously differentiable Lyapunov functions, one for each set in a particular partition of the domain determined by the projection operator of the dynamics. Interestingly, the identified Lyapunov function is the sum of two previously known and independently considered LaSalle functions. When the saddle function takes the form of the Lagrangian of an equality constrained optimization, then no projection is present. In such scenarios, if the saddle function satisfies global strong convexityconcavity, our fourth contribution establishes inputtostate stability (ISS) of the dynamics with respect to the saddle point by providing an ISS Lyapunov function. Our last contribution uses this function to design an opportunistic statetriggered implementation of the saddlepoint dynamics. We show that the trajectories of this discretetime system converge asymptotically to the saddle points and that executions are Zenofree, i.e., that the difference between any two consecutive triggering times is lower bounded by a common positive quantity. Examples illustrate our results.
2 Preliminaries
This section introduces our notation and preliminary notions on convexconcave functions, discontinuous dynamical systems, and inputtostate stability.
2.1 Notation
Let , , and denote the set of real, nonnegative real, and natural numbers, respectively. We let denote the norm on and the respective induced norm on . Given , denotes the th component of , and denotes for . For vectors and , the vector denotes their concatenation. For and , we let
For vectors and , denotes the vector whose th component is , for . Given a set , we denote by , , and its closure, interior, and cardinality, respectively. The distance of a point to the set in norm is . The projection of onto a closed set is defined as the set . When is also convex, is a singleton for any . For a matrix , we use , , , and to denote that is positive semidefinite, positive definite, negative semidefinite, and negative definite, respectively. For a symmetric matrix , and denote the minimum and maximum eigenvalue of . For a realvalued function , , we denote by and the column vector of partial derivatives of with respect to the first and second arguments, respectively. Higherorder derivatives follow the convention , , and so on. A function is class if it is continuous, strictly increasing, and . The set of unbounded class functions are called functions. A function is class if for any , is class and for any , is continuous, decreasing with as .
2.2 Saddle points and convexconcave functions
Here, we review notions of convexity, concavity, and saddle points from [21]. A function is convex if
for all (where is a convex domain) and all . A convex differentiable satisfies the following firstorder convexity condition
for all . A twice differentiable function is locally strongly convex at if is convex and for some (note that this is equivalent to having in a neighborhood of ). Moreover, a twice differentiable is strongly convex if for all for some . A function is concave, locally strongly concave, or strongly concave if is convex, locally strongly convex, or strongly convex, respectively. A function is convexconcave (on ) if, given any point , is convex and is concave. When the space is clear from the context, we refer to this property as being convexconcave in . A point is a saddle point of on the set if , for all and . The set of saddle points of a convexconcave function is convex. The function is locally strongly convexconcave at a saddle point if it is convexconcave and either or for some . Finally, is globally strongly convexconcave if it is convexconcave and either is strongly convex for all or is strongly concave for all .
2.3 Discontinuous dynamical systems
Here we present notions of discontinuous dynamical systems [22, 23]. Let be Lebesgue measurable and locally bounded. Consider the differential equation
(1) 
A map is a (Caratheodory) solution of (1) on the interval if it is absolutely continuous on and satisfies almost everywhere in . We use the terms solution and trajectory interchangeably. A set is invariant under (1) if every solution starting in remains in . For a solution of (1) defined on the time interval , the omegalimit set is defined by
If the solution is bounded, then by the BolzanoWeierstrass theorem [24, p. 33]. Given a continuously differentiable function , the Lie derivative of along (1) at is . The next result is a simplified version of [22, Proposition 3].
Proposition 2.1
(Invariance principle for discontinuous Caratheodory systems): Let be compact and invariant. Assume that, for each point , there exists a unique solution of (1) starting at and that its omegalimit set is invariant too. Let be a continuously differentiable map such that for all . Then, any solution of (1) starting at converges to the largest invariant set in .
2.4 Inputtostate stability
Here, we review the notion of inputtostate stability (ISS) following [25]. Consider a system
(2) 
where is the state, is the input that is measurable and locally essentially bounded, and is locally Lipschitz. Assume that starting from any point in , the trajectory of (2) is defined on for any given control. Let be the set of equilibrium points of the unforced system. Then, the system (2) is inputtostate stable (ISS) with respect to if there exists and such that each trajectory of (2) satisfies
for all , where is the essential supremum (see [24, p. 185] for the definition) of . This notion captures the graceful degradation of the asymptotic convergence properties of the unforced system as the size of the disturbance input grows. One convenient way of showing ISS is by finding an ISSLyapunov function. An ISSLyapunov function with respect to the set for system (2) is a differentiable function such that

there exist such that for all ,
(3) 
there exists a continuous, positive definite function and such that
(4) for all , for which .
Proposition 2.2
(ISSLyapunov function implies ISS): If (2) admits an ISSLyapunov function, then it is ISS.
3 Problem statement
In this section, we provide a formal statement of the problem of interest. Consider a twice continuously differentiable function , , which we refer to as saddle function. With the notation of Section 2.2, we set and , and assume that is convexconcave on . Let denote its (nonempty) set of saddle points. We define the projected saddlepoint dynamics for as
(5a)  
(5b)  
(5c) 
When convenient, we use the map to refer to the dynamics (5). Note that the domain is invariant under (this follows from the definition of the projection operator) and its set of equilibrium points precisely corresponds to (this follows from the defining property of saddle points and the firstorder condition for convexityconcavity of ). Thus, a saddle point satisfies
(6a)  
(6b) 
Our interest in the dynamics (5) is motivated by two bodies of work in the literature: one that analyzes primaldual dynamics, corresponding to (5a) together with (5b), for solving inequality constrained network optimization problems, see e.g., [3, 5, 14, 11]; and the other one analyzing saddlepoint dynamics, corresponding to (5a) together with (5c), for solving equality constrained problems and finding Nash equilibrium of zerosum games, see e.g., [19] and references therein. By considering (5a)(5c) together, we aim to unify these lines of work. Below we explain further the significance of the dynamics in solving specific network optimization problems.
Remark 3.1
(Motivating examples): Consider the following constrained convex optimization problem
where and are convex continuously differentiable functions, , and . Under zero duality gap, saddle points of the associated Lagrangian correspond to the primaldual optimizers of the problem. This observation motivates the search for the saddle points of the Lagrangian, which can be done via the projected saddlepoint dynamics (5). In many network optimization problems, is the summation of individual costs of agents and the constraints, defined by and , are such that each of its components is computable by one agent interacting with its neighbors. This structure renders the projected saddlepoint dynamics of the Lagrangian implementable in a distributed manner. Motivated by this, the dynamics is widespread in network optimization scenarios. For example, in optimal dispatch of power generators [11, 12, 13, 14], the objective function is the sum of the individual cost function of each generator, the inequalities consist of generator capacity constraints and line limits, and the equality encodes the power balance at each bus. In congestion control of communication networks [4, 26, 5], the cost function is the summation of the negative of the utility of the communicated data, the inequalities define constraints on channel capacities, and equalities encode the data balance at each node.
Our main objectives are to identify conditions that guarantee that the set of saddle points is globally asymptotically stable under the dynamics (5) and formally characterize the robustness properties using the concept of inputtostate stability. The rest of the paper is structured as follows. Section 4 investigates novel conditions that guarantee global asymptotic convergence relying on LaSalletype arguments. Section 5 instead identifies a strict Lyapunov function for constrained convex optimization problems. This finding allows us in Section 6 to go beyond convergence guarantees and explore the robustness properties of the saddlepoint dynamics.
4 Local properties of the saddle function imply global convergence
Our first result of this section provides a novel characterization of the omegalimit set of the trajectories of the projected saddlepoint dynamics (5).
Proposition 4.1
(Characterization of the omegalimit set of solutions of ): Given a twice continuously differentiable, convexconcave function , each point in the set is stable under the projected saddlepoint dynamics and the omegalimit set of every solution is contained in the largest invariant set in , where
(7) 
and
(8) 
The proof follows from the application of the LaSalle Invariance Principle for discontinuous Caratheodory systems (cf. Proposition 2.1). Let and be defined as
(9) 
The Lie derivative of along (5) is
(10) 
where the last inequality follows from the fact that for each . Indeed if , then and if , then and which implies that . Next, denoting and , we simplify the above inequality as
where (a) follows from the fundamental theorem of calculus using the notation and and recalling from (6) that and ; (b) follows from the definition of using ; and (c) follows from the fact that is negative semidefinite. Now using this fact that is nonpositive at any point, one can deduce, see e.g. [20, Lemma 4.24.4], that starting from any point a unique trajectory of exists, is contained in the compact set at all times, and its omegalimit set is invariant. These facts imply that the hypotheses of Proposition 2.1 hold and so, we deduce that the solutions of the dynamics converge to the largest invariant set where the Lie derivative is zero, that is, the set
(11) 
Finally, since was chosen arbitrary, we get that the solutions converge to the largest invariant set contained in , concluding the proof.
Note that the proof of Proposition 4.1 shows that the Lie derivative of the function is negative, but not strictly negative, outside the set . From Proposition 4.1 and the definition (4.1), we deduce that if a point belongs to the omegalimit set (and is not a saddle point), then the line integral of the Hessian block matrix (4.1) from the any saddle point to cannot be full rank. Elaborating further,

if is full rank at a saddle point and if the point belongs to the omegalimit set, then , and

if is full rank at a saddle point , then .
These properties are used in the next result which shows that local strong convexityconcavity at a saddle point together with global convexityconcavity of the saddle function are enough to guarantee global convergence.proving Theorem 4.2.
Theorem 4.2
(Global asymptotic stability of the set of saddle points under ): Given a twice continuously differentiable, convexconcave function which is locally strongly convexconcave at a saddle point, the set is globally asymptotically stable under the projected saddlepoint dynamics and the convergence of trajectories is to a point.
Our proof proceeds by characterizing the set defined in (4.1). Let be a saddle point at which is locally strongly convexconcave. Without loss of generality, assume that (the case of negative definiteness of the other Hessian block can be reasoned analogously). Let (recall the definition of this set in (4)). Since and is twice continuously differentiable, we have that is positive definite in a neighborhood of and so
where , , and . Therefore, by definition of , it follows that and so, . From Proposition 4.1 the trajectories of converge to the largest invariant set contained in . To characterize this set, let and be a trajectory of that is contained in and hence in . From (10), we get
(12) 
where in the second inequality we have used the firstorder convexity and concavity property of the maps and . Now since , using the above inequality, we get for all . Thus, for all , which yields
Note that both terms in the above expression are nonnegative and so, we get and for all . In particular, this holds at and so, , and we conclude . Hence is globally asymptotically stable. Combining this with the fact that individual saddle points are stable, one deduces the pointwise convergence of trajectories along the same lines as in [27, Corollary 5.2].
A closer look at the proof of the above result reveals that the same conclusion also holds under milder conditions on the saddle function. In particular, need only be twice continuously differentiable in a neighborhood of the saddle point and the local strong convexityconcavity can be relaxed to a condition on the line integral of Hessian blocks of . We state next this stronger result.
Theorem 4.3
(Global asymptotic stability of the set of saddle points under ): Let be convexconcave and continuously differentiable with locally Lipschitz gradient. Suppose there is a saddle point and a neighborhood of this point such that is twice continuously differentiable on and either of the following holds

for all ,

for all ,
where are given in (4.1). Then, is globally asymptotically stable under the projected saddlepoint dynamics and the convergence of trajectories is to a point.
We omit the proof of this result for space reasons: the argument is analogous to the proof of Theorem 4.2, where one replaces the integral of Hessian blocks by the integral of generalized Hessian blocks (see [28, Chapter 2] for the definition of the latter), as the function is not twice continuously differentiable everywhere.
Example 4.4
(Illustration of global asymptotic convergence): Consider given as
(13) 
where
Note that is convexconcave on and . Also, is continuously differentiable on the entire domain and its gradient is locally Lipschitz. Finally, is twice continuously differentiable on the neighborhood of the saddle point and hypothesis (i) of Theorem 4.3 holds on . Therefore, we conclude from Theorem 4.3 that the trajectories of the projected saddlepoint dynamics of converge globally asymptotically to the saddle point . Figure 1 shows an execution.
Remark 4.5
(Comparison with the literature): Theorems 4.2 and 4.3 complement the available results in the literature concerning the asymptotic convergence properties of saddlepoint [3, 19, 17] and primaldual dynamics [5, 20]. The former dynamics corresponds to (5) when the variable is absent and the later to (5) when the variable is absent. For both saddlepoint and primaldual dynamics, existing global asymptotic stability results require assumptions on the global properties of , in addition to the global convexityconcavity of , such as global strong convexityconcavity [3], global strict convexityconcavity, and its generalizations [19]. In contrast, the novelty of our results lies in establishing that certain local properties of the saddle function are enough to guarantee global asymptotic convergence.
5 Lyapunov function for constrained convex optimization problems
Our discussion above has established the global asymptotic stability of the set of saddle points resorting to LaSalletype arguments (because the function defined in (9) is not a strict Lyapunov function). In this section, we identify instead a strict Lyapunov function for the projected saddlepoint dynamics when the saddle function corresponds to the Lagrangian of a constrained optimization problem, cf. Remark 3.1. The relevance of this result stems from two facts. On the one hand, the projected saddlepoint dynamics has been employed profusely to solve network optimization problems. On the other hand, although the conclusions on the asymptotic convergence of this dynamics that can be obtained with the identified Lyapunov function are the same as in the previous section, having a Lyapunov function available is advantageous for a number of reasons, including the study of robustness against disturbances, the characterization of the algorithm convergence rate, or as a design tool for developing opportunistic statetriggered implementations. We come back to this point in Section 6 below.
Theorem 5.1
(Lyapunov function for ): Let be defined as
(14) 
where is strongly convex, twice continuously differentiable, is convex, twice continuously differentiable, , and . For each , define the index set of active constraints
Then, the function ,
is nonnegative everywhere in its domain and if and only if . Moreover, for any trajectory of , the map

is differentiable almost everywhere and if for some , then provided the derivative exists. Furthermore, for any sequence of times such that and exists for every , we have ,

is rightcontinuous and at any point of discontinuity , we have .
As a consequence, is globally asymptotically stable under and convergence of trajectories is to a point.
We start by partitioning the domain based on the active constraints. Let and
Note that for , , we have . Moreover,
For each , define the function
(15) 
These functions will be used later for analyzing the evolution of . Consider a trajectory of