a smooth penalty function algorithm for network-structured problems

17
220 European Journal of Operational Research 83 (1995) 220-236 North-Holland Theory and Methodology A smooth penalty function algorithm for network-structured problems Stavros A. Zenios Decision Sciences Department, University of Pennsylvania, Philadelphia, PA 19104, USA Department of Business Administration, University of Cyprus, Nicosia, Cyprus Mustafa ~ Plnar Department of Numerical Analysis, Technical University of Denmark, 2800 Lyngby, Denmark Ron S. Dembo Faculty of Management, University of Toronto, Toronto, Ont., Canada M6G 1A5 Abstract: We discuss the design and implementation of an algorithm for the solution of large scale optimization problems with embedded network structures. The algorithm uses a linear-quadratic penalty (LQP) function to eliminate the side constraints and produces a differentiable, but non-separable, problem. A simplicial decomposition is subsequently used to decompose the problem into a sequence of linear network problems. Numerical issues and implementation details are also discussed. The algorithm is particularly suitable for vector architectures and was implemented on a CRAY Y-MP. We report very promising numerical results with a set of large linear multicommodity network flow problems drawn from a military planning application. Keywords: Multicommodity networks; Nonlinear programming; Large-scale optimization; Penalty meth- ods 1. Introduction There exist many areas of application which require the solution of optimization models with tens of thousands of variables. Even with the recent developments in mathematical program- ming motivated by Karmarkar's interior point Correspondence to: Dr. S.A. Zenios, Decision Sciences De- partment, Universityof Pennsylvania,Philadelphia, PA 19104, USA. method, Karmarkar (1984) or Marsten et al. (1990), or with advances in supercomputing tech- nology, Meyer and Zenios (1988), such large-scale problems defy solution with general purpose soft- ware. It is, therefore, essential to design algo- rithms that exploit the structure of the problem. When the underlying structure is pervasive in several applications such an approach is not only very effective, but is also well justified. The optimization of network structures has progressed significantly over the last twenty years 0377-2217/95/$09.50 © 1995 - Elsevier Science Publishers B.V. All rights reserved

Upload: stavros-a-zenios

Post on 21-Jun-2016

217 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A smooth penalty function algorithm for network-structured problems

220 European Journal of Operational Research 83 (1995) 220-236 North-Holland

Theory and Methodology

A smooth penalty function algorithm for network-structured problems

Stavros A . Zenios Decision Sciences Department, University of Pennsylvania, Philadelphia, PA 19104, USA Department of Business Administration, University of Cyprus, Nicosia, Cyprus

Mustafa ~ Plnar Department of Numerical Analysis, Technical University of Denmark, 2800 Lyngby, Denmark

Ron S. Dembo Faculty of Management, University of Toronto, Toronto, Ont., Canada M6G 1A5

Abstract: We discuss the design and implementation of an algorithm for the solution of large scale optimization problems with embedded network structures. The algorithm uses a linear-quadratic penalty (LQP) function to eliminate the side constraints and produces a differentiable, but non-separable, problem. A simplicial decomposition is subsequently used to decompose the problem into a sequence of linear network problems. Numerical issues and implementation details are also discussed. The algorithm is particularly suitable for vector architectures and was implemented on a CRAY Y-MP. We report very promising numerical results with a set of large linear multicommodity network flow problems drawn from a military planning application.

Keywords: Multicommodity networks; Nonlinear programming; Large-scale optimization; Penalty meth- ods

1. Introduction

There exist many areas of application which require the solution of optimization models with tens of thousands of variables. Even with the recent developments in mathematical program- ming motivated by Karmarkar's interior point

Correspondence to: Dr. S.A. Zenios, Decision Sciences De- partment, University of Pennsylvania, Philadelphia, PA 19104, USA.

method, Karmarkar (1984) or Marsten et al. (1990), or with advances in supercomputing tech- nology, Meyer and Zenios (1988), such large-scale problems defy solution with general purpose soft- ware. It is, therefore, essential to design algo- rithms that exploit the structure of the problem. When the underlying structure is pervasive in several applications such an approach is not only very effective, but is also well justified.

The optimization of network structures has progressed significantly over the last twenty years

0377-2217/95/$09.50 © 1995 - Elsevier Science Publishers B.V. All rights reserved

Page 2: A smooth penalty function algorithm for network-structured problems

S.4. Zenios et al. / A smooth penalty function algorithm for network structured 221

with the design and implementation of special- ized algorithms. A recent survey of the state-of- the-art in this field is given in Dembo, Mulvey and Zenios (1989). Network optimization prob- lems with thousands of variables can be solved routinely on personal computers. Using parallel computing technology, researchers can solve problems with hundreds of thousands - or even millions - of variables.

In this paper we describe the design of an algorithm for solving large scale linear programs with embedded network structures. The most widely studied problem in this class is the multi- commodity network flow problem, surveyed in Assad (1987) and Kennington (1978). A related class of problems that received increasing atten- tion recently is the stochastic programming prob- lem with network recourse; see Mulvey and Vladimirou (1991) or Nielsen and Zenios (1993). It has also been revealed, due to the study of Bixby and Fourer (1988), that several well-known linear programming models have large embedded network substructures.

Our design revolves around two main ideas. First we use a penalty function to eliminate non- network constraints from the constraint set. Sec- ond, we use a simplicial decomposition algorithm to solve the penalty problem, and hence induce separability in the objective function. For prob- lems with replicated network structures - like multicommodity network flow problems or stochastic programs with network recourse - this design also induces separability of the constraint set and can be implemented on parallel comput- ers.

The idea of using penalty function methods to simplify optimization problems is probably as old as the field of nonlinear programming itself. Starting with the work of Fiacco and McCormick (1968) on penalty functions, attempts were made to solve general purpose nonlinear programming problems by reducing the problem to a sequence of problems with either no constraints or with simple constraints. Improvements on this work came later in the form of augmented Lagrangian and exact penalty functions which were explored by a number of authors (for a detailed account of the subject refer to the book by Bertsekas (1982)).

The method we propose here, however, is novel and is based on smoothed exact-penalty functions which we will refer to as e-exact penalty func-

tions. They combine the best of exact-penalty functions and augmented Lagrangians and ap- pear to be well suited to large-scale program- ming. Smoothing nondifferentiable exact penalty functions has been proposed by Bertsekas (1975). Zang (1981) used a smoothing technique to solve discontinuous unconstrained optimization prob- lems.

The simplicial decomposition algorithm that we use in order to induce separability was first proposed in Holloway (1974) as an extension to the linearization technique of Frank-Wolfe. Sig- nificant enhancements were added by Von Ho- henbalken (1977), a memory-efficient variant was developed by Hearn, Lawphongpanich and Ven- tura (1987), and an inexact variant was developed in Mulvey, Zenios and Ahlfeld (1990) where the algorithm was specialized for network structures. The use of simplicial decomposition as a device for inducing separability of large-scale problems appears to be novel in this paper.

A related line of research based on barrier methods and a trust region algorithm is reported in Schultz and Meyer (1991). Their design has been proven very successful in solving some very large multicommodity network flow problems. Another approach related to our penalty method is given by Ruszcynski (1989), who combined aug- mented Lagrangian, simplicial decomposition and dual ascent ideas; we briefly discuss its relation- ship to our method in Section 2.

This paper is organized as follows. Section 2 develops the algorithm. Section 3 provides infor- mation on key implementation aspects, like the linesearch procedure, the update of penalty pa- rameters and the linear algebra computations. Section 4 presents a summary of numerical re- sults with our implementation of the algorithm. Comparisons with alternative solution methods, like interior point algorithms, for a large-scale application are given in Section 5. Concluding remarks are given in Section 6.

2. The linear-quadratic penalty algorithm for multicommodity flows

2.1. Problem definition

We consider the following nonlinear program:

Page 3: A smooth penalty function algorithm for network-structured problems

222 S.A. Zenios et al. / A smooth penalty function algorithm for network structured

(NLP)

minimize f ( x ) x

s . t . A x = b , Ex < d , l <_x <_u,

where: f : ~ n ~ ~ is the objective function, assumed to be convex and continuously differentiable. x E ~n is the vector of decision variables which represent flows on a graph. A is an m × n constraint matrix with network structure (i.e., it could be the node -a rc incidence matrix of a network flow problem, or it could be a block-diagonal matrix where each block is a node -a rc incidence matrix as occurs in multicom- modity network flow, stochastic network flow and time staged problems). We assume throughout this paper that

A = diag[A1, A 2 . . . . . AK] ,

where K > I , and each At, t = l . . . . . K, is a node -a rc incidence matrix. E is the s × n matrix of side (i.e., non-network) constraints. l, u ~ R ~ are bounds on the variables. b ~ ~ ' , d ~ ~ are the right-hand side coeffi- cients of the constraints. Also, let

X = { x l A x = b , l < x < u }

and

~ = { x l x ~ X , Ex < d } .

Throughout the manuscript, transposition is indi- cated by a superscript T and Vf denotes the gradient vector of the function f .

2.2. Linear-quadratic penalty functions

We present first a preview of the main compo- nents of the algorithm. The first design decision is to remove the side constraints and append them to the objective using an exact penalty func- tion

pi( ti) = max{O, tj} (1)

where t i = E i x - d i. Thus by placing the side constraints into the objective function using the penalty function (1), we obtain a problem with

network constraints. In particular, for multicom- modity flows or stochastic networks, the penalty problem has a Cartesian product constraint set. Unfortunately, using the exact penalty function (1) produces a non-differentiable problem. In ex- periments with exact penalty functions for linear network problems, Gamble et al. (1991) indicate that only modest performance improvements over direct application of the simplex method are pos- sible.

To avoid the difficulties of non-differentiabil- ity we make the second design decision: smooth the exact penalty. This approach has been pro- posed in the context of discontinuous optimiza- tion by Zang (1981). Bertsekas (1975) proposed smoothing in the context of exact penalty func- tions. For the exact penalty function (1), we con- sider the following smoothed penalty functions:

0 if t_<0, t 2

$ 1 ( e , t ) = 2e i f 0 < t < e , (2)

t - ½ e i f t > e ,

l 0 if t < 0 , t 4 t 3

¢ 2 ( e , t ) = | - 2 e - - - ~ + ~ i f0_<t_<e , (3)

~ t - ½e if t < e ,

where t is a scalar real variable. ¢1(e, t) is once continuously differentiable, while ¢2(e, t) is twice continuously differentiable. Using ¢2(e, t) as the smoothing function would allow us to use second order algorithms, such as the truncated Newton method, Dembo (1987), for the solution of the penalized problem. However, with second order methods the penalty function would be nonsepa- table. Potentially decomposable constraint sets would be coupled through the objective. Hence we make the third design decision: solve the penalty problems using a linearization technique, such as simplicial decomposition. In this case we may use thm(e, t) as our smoothing function. An additional advantage of thl(e, t) over $2(e, t) i s that it is more stable numerically as e becomes very small. ¢1(e, t) is illustrated in Figure 1.

These ideas are made precise in the following sections. We first outline the basic ideas underly- ing the notion of e-optimality. We then proceed with a description of the linear-quadratic penalty algorithm. We would like to point out that several

Page 4: A smooth penalty function algorithm for network-structured problems

S~4. Zenios et a L / A smooth penalty function algorithm for network structured 223

Pt 6 Figure 1. The linear-quadratic smooth penalty function

variations are possible around the general design outlined above that are worth additional investi- gation. For example, we may use successive linear programming to solve the penalty problem, again inducing separability or use a truncated Newton method with the block-partitioning techniques of Zenios and Plnar (1992) to exploit separability.

2.3. Aproximate optimality and e-exactness

The linear-quadratic penalty function viewed as an approximation to the gl penalty function possesses some approximate exactness properties, established in Plnar and Zenios (1994). We give a summary of the basic ideas here.

We consider the following smoothed penalty version of the nonlinear program (NLP).

(PNLP)

min~P~, ~ ( x ) = f ( x ) + ~ k 4'a(e, t j) , x ~ X " j = l

where

t j = E j x - d j for j = 1, 2 , . . . , s ,

and ~ is a positive real number. In this section we investigate the relationship between an opti- mal solution to the original problem (NLP) and an optimal solution to the e-exact penalty prob- lem (PNLP). Associate the multipliers: A with

constraints Ax = b, to > 0 with constraints Ex < d, and OU, 0 t > 0 with constraints l < x < u, and let g = V f .

The first-order optimality conditions for (NLP) are then

g ( x ) +Aa'A +ETto + 0 u - 0 l = O, (4)

Ax = b, (5)

Ex <d, (6)

l <x <u, (7)

W<___O, O u ~ O , o l e o , (8)

w(Ex - d) = 0, (9)

ou(x - u ) = 0, ( 1 0 )

O'(x - l) = O. (11)

If we assume strict complementarity and sec- ond-order sufficiency then x*, to*, h*, 0* satisfy- ing the above are unique. We compare these conditions with the optimality conditions of the e-exact penalty problem (PNLP). Associating A~ with constraints Ax = b and 0~, 0~ >_ 0 with con- straints l _< x _< u in (PNLP), we have

g ( x ) dr'U ~ V~I(E , t j ) "[-ATI~e"[- Off --Ül e = O, y=l

(12)

= b , (13)

l <x <_u, (14)

o ~ ( x - u ) = o, ( 1 5 )

O~(x - l) = O, (16)

o~ >__ o, o~ >__ o. (17)

Letting

d ( x ) = { j lO<t j_<e}

be the set of active constraints and

7/'( x ) = {j l ty > e}

be the set of violated constraints, the above con- dition (12) reduces to

g(xe) + ~ E ( E j x ~ - d j ) E T + I ~ E ET e j~(x~) j~7/(x~)

+ATA~ + 0~ + 0~ = 0, (18)

Page 5: A smooth penalty function algorithm for network-structured problems

224 S~4. Zenios et al. / A smooth penalty function algorithm for network structured

where x~ denotes a solution to the optimality conditions of the penalized problem (PNLP). Now let

i/x x~ for j ~ ~(x~), ( , , , j j -dj) = for j ~ ¢ ( x ~ ) , (19)

otherwise.

Then the above reduces to

2.4. The Linear-Quadratic Penalty (LQP) algo- rithm

The use of smoothed exact penalty to elimi- nate side constraints motivates the following algo- rithmic framework:

The Linear-Quadratic Penalty Algorithm. Step O. (Initialization.) Find an initial feasible

solution for the relaxed problem

g(x~) + EToj~ +Aa'A~ + Off - O~ = O,

which is reminiscent of the Kuhn-Tucke r condi- tion (4) for optimality in the original problem. Thus a solution x~ to the e-exact penalty prob- lem and the multipliers toe, As, Off, 0~ defined above, satisfy conditions (4), (5), (7), (8), (10), and (11) of the conditions for optimality of the origi- nal problem. Optimality for the problem (NLP) would be assured if we had primal feasibility (i.e. Ex~ < d, see (6)) and complementary slackness (i.e. oJ~(E% - d) = 0; see (9)).

In solving (PNLP) we compute an x~ that satisfies (13)-(17), that is e-feasible, i.e. ~r(x~) = ¢ (and hence E% - d < e) and that (almost) satis- fies complementary slackness (9). Since, by con- struction, tog > 0, a necessary condition for e-op- timality is

Ex~ - d < e (e-feasibility),

which implies (by definition of w~)

w,( Ex~ - d) < tze (e-complementary slackness).

Thus, for e small enough, tze will be small. Such a solution will be termed e-optimal for (NLP). The above analysis gives us insight as to how two of the key parameters in an algorithm for (PNLP) may be adjusted from iteration to iteration. First, e should start out relatively large and should be reduced so that /ze becomes an acceptable com- plementary slackness error in the limit. If no constraint is violated (i.e. ~ ( % ) = ¢) then leave /~ unchanged. If at least one constraint is violated increase/x according to the formula

- dj) ) ,---max - .

e

(RNLP)

minimize f ( x ) x

s.t. Ax = b l <x <u

Set k = O, let x ° be the optimal solution of this problem. If x ° ~ 0 stop. Otherwise choose/x ° > O, e ° > O, go to Step 1.

Step 1. Using the violation tj = (Ex - d)j for all j as the starting point for evaluating the penalty function, solve the problem

mincb~,,(x) = f ( x ) + tz ~ ~bl(e, tj). x E X j = l

Let x *~ denote the optimal solution. Step Z If x*k satisfies termination criteria ter-

minate. Otherwise, let x k + l ~ x *k, update the penalty parameters ~ and e, set k ~ k + 1 and proceed from Step 1.

Details on this general algorithmic framework are completed in the following sections - see also Figure 2.

Most of the computational effort of the algo- rithm is in solving the smoothed penalty problem (PNLP) in Step 1. This problem is once continu- ously differentiable, and can be solved using spe- cialized nonlinear network flow algorithms such as the truncated Newton described in Dembo (1987). However, the penalty function ~b(e, t) is non-separable in x (recall that tj = ( E x - d)j). Hence, the fact that the original problem (NLP) is such that X is decomposable, (i.e. A = diag[A 1, A z . . . . . AK], where K > 1) provides the incentive to use an algorithm for (PNLP) that is based on a linearization of the smoothed exact penalty function. In this way the algorithm de- composes. Examples of such algorithms are the

Page 6: A smooth penalty function algorithm for network-structured problems

S.A. Zenios et aL / A smooth penalty function algorithm for network structured 225

MAJOR ITERATION

INITIAL FEASIBLE POINT FOR EACH COMMODITY

YES

SOL ?L~ON

k ~ k + l

L UP°ATE PE AL I PAR =RS

1 COMPOTE PENALTY

FUNCTION

o~,ax ~) = f(~) + uC,(x k) SIMPL1CIAL DECOMPOSITION

ALGORITHM

GENERATE DESCENT DIRECTION BY SOLVING L1NEARIZED

SUBPROBLEM FOR EACH COMMODITY

MINOR

SOLVE NONLINEAR MASTER ITERATION PROBLEM OVER ALL COMMODITIES

C. m

NO

ES I Y

Figure 2. A flow diagram of the linear-quadratic penalty algorithm

The Simplicial Decomposition Algorithm. Step O. Set v = O, and use z ° ~ x k ~ X a s the

starting point. Let Y = ¢, and v ~ 0 denote the set of generated vertices and its cardinality, re- spectively.

Step 1. (Linearized subproblem). Compute the gradient of the penalty function abel ~ at the cur- rent i terate z ~ and solve a linear program to get a new vertex of the constraint set, i.e,, solve for

U* = Argmin y T V ~ , ~ ( z v) y ~ X

and let Y = Y U {y*}, v ~ v + 1. Step 2. (Nonlinear master problem.) Using the

set of vertices Y to represent a simplex over the constraint set X, find an optimizer of the penal- ized objective function ~,,~ over this simplicial subset of X. Let

w * = Arg min qO~,#(Bw) wE W v

where

Wv= W i w i = l , wi>_O, V i = l , 2 . . . . . v i=l

and B = [ya[ y21 . . . I yq is the basis for the sim- plex generated by the set of vertices Y. The optimizer of q~,~ over the simplex is given by z v+l =Bw*.

Step 3: Let u ~ u + 1, and return to Step 1.

network specialized reduced gradient method de- scribed in Dembo and Klincewicz (1981), succes- sive linear programming, and simplicial decompo- sition. We solve (PNLP) using simplicial decom- position, see Von Hohenbalken (1977), Hearn , Lawphongpanich and Ventura (1987) and Mul- vey, Zenios and Ahlfeld (1990).

2.5. Linearization via simplicial decomposition

Simplicial decomposit ion iterates by solving a sequence of linear problems to generate vertices of the polytope X. A nonlinear master problem optimizes the penalized objective function q~,~ on the simplex specified by the vertices generated by the subproblems. The simplicial decomposi- tion algorithm for (PNLP) at the k-th i teration of the LQP algorithm may be stated as follows:

At Step 1 the algorithm solves a linear approxi- mation to the nonlinear program. I f the con- straint matrix has a block diagonal structure the problem can be solved independently for each block of the constraint matrix A:

Step 1 (Decomposed linear subproblems). For each t -- 1, 2 , . . . , K solve

Minimize yTVttlb ,e( z v)

s.t. AtY t = bt,

It <-Yt <- ut.

We use t to index the t-th block of the constraint matrix A = diag[A1, A 2 , . . . , A K ] . Similarily, Yt, bt, l t , u t , Vtc]9 denote subvectors or the respective vectors, corresponding to the variables of the t-th block. The subproblems are linear network pro-

Page 7: A smooth penalty function algorithm for network-structured problems

226 S.4. Zenios et al. / A smooth penalty function algorithm for network structured

grams for each commodity and thus may be solved in parallel.

The nonlinear master problem in Step 2 is of much smaller size than the original (NLP). Typi- cally, the number of vertices upon termination of the algorithm does not exceed 100. Furthermore, it has a simple structure: a simplex equality con- straint and bounds on the variables. However, this problem is likely to present difficulties due to the poor scaling of the objective function as gets larger. Discussions on the efficient solution of the master problem are deferred until Section 3.3.

2.5,1. Termination criteria in simplicial decomposi- tion

The termination criteria used in the simplicial decomposition phase are based On monitoring the duality gap, the number of simplicial decomposi- tion iterations, i.e., the number of linearized sub- problems, and the change in the objective func- tion value in successive simplicial decomposition iterations.

Consider iteration i of the penalty algorithm. Let x i and (/)i~,~ denote the minimizer of q~u,~ and the minimum at x i respectively. The duality gap is computed as

U ~ _ L ~ ; egap U v

where U ~ and L ~ are upper and lower bounds to the optimal objective function value ~ , ~ at itera- tion v of the simplicial decomposition algorithm. The lower bound is computed using a first order Taylor series expansion o f the function @,,~ around the current iterate z ~. The objective func- tion value provides an upper bound since we are minimizing. The change in objective function value at iteration v of the simplicial decomposi- tion algorithm is computed as

eob j = ~b ,~(z~_l) (21)

The maximum number of simplicial decompo- sition iterations is set to be 20 in our implementa- tion. The tolerance for egap is 10 -3, and for eob i is 10 -6. The simplicial decomposition algorithm ter- minates whe n egap ~< 10 -3, o r Cob j ~ 10 -6, or v > 20. This is followed by an update of the penalty parameter/.~ and the smoothing parameter e. If

the termination criteria for the penalty algorithm are not met, the process restarts.

2.6. Extensions to problems with coupling variables

The LQP algorithm can be extended to solve problems of the form

(NLPSV)

x,y Minimize f ( x , y)

s.t. Ax = b , Ex + S y < d ,

I x <_X <_ Ux,

l y ~ y ~ U y ,

where: y ~ R v are the side variables. E is the s × n matrix of side (i.e., non-network) constraints. S is the s × p matrix of side (i.e., non-network) constraints corresponding to coupling variables. Ix, u x ~ ~n are bounds on the network flow vari- ables x. ly, Uy ~ ~ v are bounds on the side variables y.

This extension of LQP to the solution of prob- lems with side constraints and side variables is developed in Pmar and Zenios (1993), where extensive numerical results are also given. The key issue in this extension is the handling of coupling variables. The linear-quadratic penalty function is used to penalize the constraints Ex + Sy < d. At Step 1 of the simplicial decomposition algorithm a new vertex (y ) i s generated as the solution to the following subproblem:

Minimize xTVx~( x ~, y~) + yTVy~( x ~, yV) x,y

s.t. /Ix = b, l x < x < u x, ly _<y <Uy,

where (~) is the iterate as the v-th iteration of simplicial decomposition. This problem decom- poses into two independent linear programs as follows:

Minimize X T V x ~ ( X v, y~) x

s.t. Ax = b , I x < x <_ Ux,

Page 8: A smooth penalty function algorithm for network-structured problems

S.A. Zenios et al. / A smooth penalty function algorithm for network structured 227

and

Minimize yTVyt~( X u, yV) Y

s . t . ly _< y ~ Uy.

The first problem is a linear network problem and is solved using the network simplex methods. It could also be decomposed if the constraint matrix A has a block-diagonal structure. The second problem in the side variables is solved by inspection, by assigning each component yj of the vector y to its lower or upper bound depending on the sign of the gradient Vy~(X ~, y~), i.e.,

(Uy)j if Vr~(x ~, y~) >0 , (22)

YJ = (ly)j if Vyfl~(x ~, y~) < O.

2. 7. Relation to augmented Lagrangian methods

A decomposition method based on augmented Lagrangian was introduced by Ruszcynski (1989) to solve block diagonal linear programming prob- lems. This method shares some common features with our penalty method Both algorithms treat the coupling (side) constraints implicitly. Rusz- cynski's algorithm employs an augmented La- grangian formulated based on a quadratic term while we use a linear-quadratic penalty function. The result in both cases is a nonlinear non-sep- arable network problem with a block-diagonal constraint matrix.

The LQP algorithm involves two parameters/~ and e to control the accuracy of the solution. While the penalty algorithm is not a finite algo- rithm, it possesses some approximation proper- ties that can be derived using the gl exact penalty function. Ruszcynki's algorithm is a finite algo- rithm, The LQP algorithm does not involve any multiplier iterations whereas Ruszcynski's algo- rithm uses the multiplier iteration as a dual as- cent for the Lagrangian. To solve the resulting non-separable network problem with block-diago- nal constraint matrix, both algorithms use similar, but distinct, column generation ideas the alter- nate between a subproblem and master problem phase.

Finally, w e remark that the LQP algorithm offers some flexibility in the choice of the solu- tion algorithm for the block-diagonal network

problem. In this connection, the simplicial de- composition algorithm can be replaced by other suitable algorithms without destroying the ap- proximation properties of the algorithm. This flexibility is apparently absent in Ruszcynski's algorithm. The LQP algorithm can also handle nonlinear objective functions.

3. Implementation and numerical issues

There are three components of the LQP algo- rithm that deserve special attention in the imple- mentation: 1) the scheme for estimating and up- dating the multipliers {/x k} and the tolerance {ek}; 2) the linesearch algorithm for a piecewise linear-quadratic function; and 3) the nonlinear programming algorithm for solving the uncon- strained master problem. The master problem is typically small but may be ill-conditioned. The choices made for these three components are crucial for a robust implementation.

3.1. Penalty parameter adjustments

Suppose x k ~ X , tz k, e k are given at iteration k of the LQP algorithm. Also let y~= Ex ~ - d. The iterate x k is termed e-feasible if the index set of violated constraints ~e(x k) as defined in Section 2.3 is empty.

We distinguish between the following two cases when updating the penalty parameters:

Case 1: If ~ ' (x k) = ¢, this is an indication that the magnitude of the penalty parameter /~ was adequate in the previous iteration since e-feasibil- ity is achieved. In this case the infeasibility toler- ance e may be reduced.

Case 2: If ~"(x ~) v~ ¢, the current point is not e-feasible, an indication that the penalty parame- ter/z should be increased. One possible choice is to do so proportionately to the degree of infeasi- bility. Let 3' = ~/e k be a target degree of infeasi- bility where 77 ~ (0, 1]. We consider the following update equation:

/.k+l = t z k y j y , (23)

or equivalently,

/z k+l =/z ~ YJ (24) ~Te k •

Page 9: A smooth penalty function algorithm for network-structured problems

228 S.4. Zenios et al. / A smooth penalty function algorithm for network structured

I I E a)

It

Y)

uk + lCj(e,yj)

••slope = /~+x~

I I y •

It Y)

Let y = qe

~ + 1 = inax~+l J

Figure 3. Multiplier update

Since I ~'~(xk)l > 1, we get

/ ~ k + l = /Zk max yj. (25)

"rle k ] ~ . ( x k)

In summary we have the following update pro- cedure, see also Figure 3.

Pick ~Ta, "112 ~ (0 , 1] I f ~ ' ( x k) =

e~+t = max{emin, "ql 6k} else

~ k

~ k + l = max Yi, 972 8k jE~F'(X k)

where emi n is a suitable final feasibility tolerance. A suitable initial value for /z is found through some preliminary experimentation. With the lin- ear programs used in this study, the absolute maximum of objective function coefficients proved to be a good choice. The solution to the multi- commodity network flow problem obtained by ignoring the side constraints can be used to ob- tain an initial value for e. A reasonable choice is to pick a value equal to a fraction of the maxi- mum of the side constraint violations, i.e., in the interval (0, maxj ~ ~(x0)yj). The value of parame- ters ~7~ and ~Ta was taken to be 0.5 for all compu- tational tests reported in this study.

3.2. Linesearch procedures

We consider now the problem of minimizing a piecewise linear-quadratic function along a direc- tion of descent. This problem is solved at every iteration of the master problem in simplicial de- composition. It can be posed as follows:

min cI~ ,~(y + ap) (26) O_~a_~l

where p = [Pl[ P21 "'" [ PK] T is a descent direc- tion at point y. That is, Vqbf,~ .p < 0. (How to compute such a descent direction is left for the next section.)

The function @~,, given by (2) is once continu- ously differentiable and convex, and hence there exists an optimal solution to (26) at the minimum of a quadratic segment. Hence, we may apply, in principle, any one-dimensional search algorithm such as a quadratic interpolation with safeguards, see, e.g. Gill, Murray and Wright (1981). I t is likely, however, that a linesearch based on quadratic interpolation will be inefficient when the function has a large number of linear seg- ments, and the linesearch is started far from the minimizer. The linesearch we develop at tempts to build a coarse approximation to the function based on local information. In general, when far from a minimum the function is viewed as piece- wise linear. At each stage the function values at two distinct points bracketing the minimum are assumed to be known. The left hand point, a L, has a negative gradient and the gradient at the right hand point, a g , is positive. We say that the left hand point is quadratic (linear) if i t lies on a quadratic (linear) segment. By the left hand line

Page 10: A smooth penalty function algorithm for network-structured problems

S.,4. Zenios et al. / A smooth penalty function algorithm for network structured 229

we mean the line corresponding to the linear segment when the left point is linear. Similarly, the right hand quadratic corresponds to the quadratic approximation at a right hand point that is quadratic. The linesearch algorithm has the following general f o r m :

START with (aL, teR ) bracketing the minimum. COMPUTE a new approximation to the mini- m u m ~; W H I L E V q~(y + tep) 4= 0

I f V ~ ( y + ap) > O, a n ~ dr, I f V ~ ( y + tep) < 0, te L ~ t~.

Let us now examine the three cases that may arise:

CASE L The left and right hand points are both linear. The function is approximated as a piece- wise linear function and the next approximation is taken to be the interaction of the left and right hand lines.

CASE II. The left hand point is linear and the right hand point is quadratic (or vice versa). We assess that the current left and right hand points are not in the neighborhood of a minimum. We therefore continue to t reat the problem as if it were piecewise linear and our next approximation point is the intersection of the linear approxima- tions at the two points.

CASE IlL The left and right hand points are quadratic. Here we have reason to suspect that we have bracketed the quadratic segment on which the minimum lies. We therefore fit a quadratic through the two points which will be exact in a neighborhood of the solution.

ble step length that will not violate the bounds. This problem is studied in Mulvey, Zenios and Ahlfeld (1990). Here we consider additional diffi- culties that arise due to the piecewise linear- quadratic nature of the objective function.

The master program can be written in the form

min qb~ ~(Dw) (27) w>_0 '

where

D = [Ya -Yv l Y2 -Yv [ . - - l Y v - 1 -Y~]

is the derived linear basis for the simplex gener- ated by the vertices Yl, Y2,-. . , Y~- We denote by w the vector [w,, w 2 . . . . . wv-1] and the solution for wo is computed as

v - - 1

w~ = 1 - E wi. (28) i = 1

Recall that at the current iteration we have v - 1 active vertices (i.e. w i > O, for i = 1 . . . . , v - 1) and the last vertex Yv lies along a direction of de- scent. Hence, given a point z v ~ X 0 a descent direction to (27) can be obtained as the solution to

(DTMD)p = --DTVqbt~#(Z~), (29)

where the choice of the matrix M is discussed next.

3.3.1. Orthogonally projected gradient steps If we choose M to be a conformable identity

matrix we have

3.3. Computing directions of descent

We return now to the solution of the master problem in simplicial decomposition. It is a non- linear problem with simple bounds and a single equality (simplex) constraint. There are several s tandard methods that can be used for its solu- tion. I f the simplicial decomposit ion algorithm drops vertices that carry zero weight at the opti- mal solution of the master problem, then subse- quent master programs are locally unconstrained. Hence, methods of unconstrained optimization can be used to compute a descent direction. A simple ratio test determines the maximum feasi-

(DTD)pI = - D T V ~ , ~ ( z ~) (30)

and Pl is a descent direction since DTD is posi- tive definite since D is full rank. An exact method can be used to solve for PI. Note, however, that D is of dimension n × ( v - 1). In general n is very large (more than 100000 for some of the larger test problems), whereas v starts f rom 2 and rarely exceeds 100. Hence DTD is a relatively small matrix of dimensions

( v - 1) X ( v - 1)

which can be easily inverted once it has been explicitly formed. To avoid the computat ion of

Page 11: A smooth penalty function algorithm for network-structured problems

2 3 0 SM. Zenios et aL / A smooth penalty function algorithm for network structured

the product DTD we use a QR factorization of D and hence solve

( RTR)PI = -DTVqbg,e ( z ) , (31)

where R is an upper triangular matrix. Numerical results with method, and comparisons with an LU decomposition of DTD and a conjugate gradient solver are discussed in Section 4.

3.3.2. Truncated Newton steps If we choose M to be the Hessian matrix

( H = ~72(Jg/~,e(Z)) we have

(DTHD)PTN = --DTVqbg,~(z). (32)

Since the once continuously differentiable penalty function given by (2) is used in this paper, the function does not have continuous second deriva- tives. In this case, the Hessian is not unique at points where the coupling constraints are bind- ing. This can be avoided by using the twice differ- entiable penalty function given by (3). However, this function is numerically less stable. In both cases, the system (32) is positive semidefinite which precludes the possibility of using a direct method (the Hessian matrix could have zero ele- ments along the diagonal when the penalty func- tion is at a linear segment). However, it can be solved inexactly using a conjugate gradient solver to compute a truncated Newton step.

3.3.3. Quasi-Newton steps If we choose M to be the Hessian matrix

(H~ = V2~,,,(z~)) evaluated at the solution to the master problem of the previous iteration, we have

(DTH,,D)pQN = -DTVCg, , (zV) . (33)

This approach, first suggested in Mulvey, Zenios and Ahlfeld (1990), is particularly attractive since it removes the requirement to form the product DTHD at every step of the master problem algo- rithm. Instead, we work with a fixed matrix DTH~D which is computed only once. Unfortu- nately the Hessian is not guaranteed to be posi- tive definite and the system is solved inexactly using a conjugate gradient solver.

3.4. Termination criteria

In this section we examine the termination criteria used in the LQP algorithm. We show that

it is possible to obtain lower and upper bounds on the optimal objective value of the original problem (NLP) even in the presence of inexact minimizations.

Let x* be an optimal solution of (NLP) and 2~,~ an optimal solution of (PNLP) for given penalty parameters /z and e. Then

cp~,~( ~,,~) < f ( x* ) (34)

since (PNLP) is a relaxation of (NLP). Therefore the optimal solution of (PNLP) is a lower bound for the optimal objective value. But in the pres- ence of inexact minimizations of the penalized objective function q~,,~, this is not always guaran- teed to be a lower bound. Hence, we consider the first order Taylor series expansion of q0,, around z ~ X :

= - z ) %Ax) + (x

+o(llzll z) V x e X , (35) and ignoring the second order term define the function, h~,,~ : R n ~ R,

hg,~(x) qb,~(z) -[- ( X T = - z ) V'~g,~(z). (36)

By convexity of 4~g,~,

mi r th (x ) _< 4~ . , . (~ , . ) , x ~ X

and hence,

minhg . ( x ) <f(x*) . x ~ X "

This bound is already computed by the simplicial decomposition algorithm that generates extreme points of X by minimizing a linearized approxi- mation to the objective function over X. Our computational experience shows that this lower bound may not be very tight.

We now proceed to describe the procedure for generating upper bounds in the linear-quadratic penalty algorithm. For a more general discussion on bounding exterior penal ty function algorithms see Fiacco and McCormick (1968). Define the set

R°= { x ~ X l E x <d}

and assume that R ° is non-empty. Let x ~ R ° and ~g,~ be a - perhaps approximate - optimal solution of (PNLP). Then a new interior point x is generated as follows: let y = E£~,,~- d, and ! =

Page 12: A smooth penalty function algorithm for network-structured problems

S.4. Zenios et al. / A smooth penalty function algorithm for network structured 2 3 1

{ i l y / >0} . Let y ' = E x - d and y / ' <0 , V i = 1 , . . . , s. Calculate

Yi /3 = max (37)

i ~ l Y i - - Y [

and define

x = (1 - /3)2~,~ +/3x. (38)

It can be shown that x ~ / 2 and provides an upper bound, Fiacco and McCormick (1968, Theorem 29, p. 107). The same result also proves that the upper bound converges monotonically to the opti- mal objective value. Obviously this procedure re- quires an interior point to be generated at the beginning of the algorithm. We adopted a heuris- tic procedure given in Zenios, Qi and Armstrong (1990) to generate a starting interior feasible so- lution for multicommodity network flow prob- lems.

We have therefore both upper and lower bounds for the optimal objective value during the execution of the algorithm. The algorithm termi- nates when both of the following error measures are within acceptable tolerance.

1. Absolute error in side constraint feasibility:

[I Ex - d [I = < e m i n •

2. Bound gap:

f ( x ) -hg ,~(x ) h, ,~(x) < egap,

where x is the current iterate and x is obtained from (38).

4. Computational results

The LQP algorithm was implemented for the multicommodity network flow structures. We used the simplicial decomposition implementation from the system GENOS of Mulvey and Zenios (1987) to solve the penalty subproblems. A main program around simplicial decomposition sets up the penalty function; the master problem solver in GENOS has been modified to handle the factorization procedures outlines in section 3.3. The code - which we call G E N O S / M C for Gen- eralized Network Optimization System for Multi-

Commodity networks - is written in F O R T R A N 77.

In this section we report summary computa- tional results. The intention is not to present an exhaustive list of experiments, but to highlight key aspects of the algorithm and illustrate its suitability in solving very large problems. We also provide comparisons with alternative solution al- gorithms for identical test problems, thus illus- trating the effectiveness of the algorithm and the efficiency of our implementation.

In all computational tests, we have e,a . = 10 -5, and egap = 10 -2 on termination. As we remarked earlier the method does not generate good lower bounds. This explains the high value of the error egap. Solving the penalized problem (PNLP) to optimality would remedy this situation at the expense of higher computation times. We point out, however, that the answers we obtained match optimal values given by MINOS to five digits, for the problems we solved with both systems.

All results reported in subsequent sections re- fer to CPU seconds on a VAX 6400 running VMS 5.3, unless otherwise indicated. The pro- gram was compiled with default optimization level. All reported times exclude inpu t /ou tpu t of data.

4.1. Test problems

We consider two classes of test problems. One class of problems was obtained from a Military Airlift Command application. They are referred to as the Patient Distribution System (PDS) prob- lems and are used to make decisions on the evacuation of patients from Europe. PDSt de- notes a problem that models a scenario of t days. They are linear multicommodity network prob- lems with eleven commodities. The size and com- plexity of the models increase with t.

The KENx problems are randomly genera ted multicommodity networks. They were generated by the code M N E T G N of Ali and Kennington (1977). The size and characteristics of all the test problems we solved from both classes are summa- rized in Table 1.

4.2. Solving the master problem

We experimented with the three methods of Section 3.3 for computing descent directions: Re-

Page 13: A smooth penalty function algorithm for network-structured problems

232 SM. Zenios et al. / A smooth penalty function algorithm for network structured

Table 1 Test problem characteristics

Test Network Formulation

problem Arcs Nodes Com- mod- ities

Linear Programming

Rows Columns

KENll 176 121 121 14694 21349 KEN13 225 169 169 28632 42659 PDS1 339 126 11 1473 3816 PDS3 1117 390 11 4593 12590 PDS5 2149 686 11 7546 23639 PDS10 4433 1399 11 15389 48763 PDS15 7203 2125 11 23375 79233 PDS20 1 0 1 1 6 2447 11 31427 105728

"~00"

WOO'

-----e----- '3ENOS I . -4=Nmn

_ near-auaGrauc Ilnesearcn

• 3 20 30

Table 2 Performance of various step selection strategies with PDS1

Step selection Linear algebra solver

LU QR Conj. grad.

Reduced gradient 319.41 318.0 244.49 Newton NA a NA 188.14 Quasi-Newton NA NA 130.4

a NA: Not applicable.

duced gradient, Newton, quasi-Newton. We also used exact solvers (LU or QR factorizations) and an iterative solver (conjugate gradient) for com- puting truncated Newton directions. The perfor- mance of the three methods is illustrated with two test problems PDS1 and PDS3 in Tables 2 and 3. The measure of performance is the total solution time given in CPU seconds. Computing Quasi-Newton steps using a conjugate gradient solver has a distinct advantage over all other methods. It is the default master program solver of GENOS/MC.

Sirapacml lteraaons

.~00'

2 0 0

100

/ el GENOS Imesearcrl / P

/ -- LJneat-ouaaratt¢ unesearcn

'0 2'0 3'0

Figure 4. Comparison of the piecewise linear/quadratic line- search with the quadratic interpolation with safeguards

4.3. Performance of linesearch procedures

The nonlinear master program can use either the piecewise linear-quadratic linesearch of Sec-

Table 3 Performance of various step selection strategies with PDS3

Step selection Linear algebra solver

LU QR Conj. grad.

Reduced gradient 3910 3924 5644 Newton NA a Na 1998.7 Quasi-Newton NA NA 1239

a NA: Not applicable

tion 3.2, or a standard quadratic interpolation linesearch with safeguards that is already in GENOS. Both linesearch routines were tested and compared on PDS1, Figure 4 compares the two linesearch routines with respect to number of function evaluations and CPU time. Major itera- tions refer to executions of Step 2 of the LQP algorithm, whereas simplicial iterations is the to- tal number of simplicial decomposition steps re- quired for the solution of penalty problems. First we observe from the results that the overall per- formance of the LQP algorithm does not depend on the choice of the linesearch procedure. This

Page 14: A smooth penalty function algorithm for network-structured problems

S.A. Zenios et al. / A smooth penalty function algorithm for network structured

Table 4 Performance of the advanced start procedure

233

Test problem Artificial basis start Advanced start

Subprob. time Master time Total Subprob. time Master time Total

PDS1 26.33 112.83 139.27 7.10 104.25 111.35 PDS3 251.65 1080.25 1331.9 59.24 929.54 988.99

was, of course, anticipated since both linesearch procedures solve each master problem to the same level of accuracy. Second, we observe f rom the results of Figure 3 that the two methods are at par with each other with a slight advantage of the G E N O S linesearch. This result is somewhat surprising since we expected the piecewise l inear-quadratic linesearch to t ake advantage of the special structure of the penalty function. For problems with a large number of linear segments and a small number of quadratic segments this should be the case. Both linesearch routines are par t of G E N O S / M C and a user controlled pa- rameter determines which one is used.

4.4. Restart procedures for the network subprob- lerns

One of the major components of the solution t ime comes f rom the effort expended in solving the subproblems in the simplicial decomposit ion algorithm. The subproblems are linear network flow problems for each commodity and can be solved with the network specialization of the sim- plex method starting with an artificial basis at each subproblem phase. Alternatively, an ad- vanced basic start procedure can be incorporated into the subproblem phase to reduce the number of pivots to optimality, This can be achieved, for instance, using the maximal basis procedure of Dembo and Klincewicz (1985). This procedure uses a greedy heuristic to construct a basis tree f rom a given feasible iterate. It is also possible to save the basis arrays of the previous iteration and

use the previous basis as the initial basis in the current subproblem phase. Both these proce- dures were implemented and tested. We illus- t rate the effect of the second advanced start procedure on two test problems, PDS1 and PDS3 in Table 4. As can be observed f rom the table, significant improvement is realized with the ad- vanced start procedure. It is used by default in the LQP algorithm.

4.5. Performance of the algorithm

In this section we summarize the performance of the algorithm on a subset of three PDS prob- lems. Table 5 provides the statistics with respect to the CPU time spent in the solution of subprob- lems and master problems of the simplicial de- composition algorithm. The number of major it- erations refers to the number of penalty mini- mizations, i.e. executions of Step 2 of the LQP algorithm. The number of vertices retained upon termination of the algorithm and the total num- ber of simplicial decomposit ion iterations are also reported.

The objective function values upon termina- tion of the algorithm for these problems agree to the first five digits with the objective function values repor ted by MINOS of Murtagh and Saun- ders (1983). We note that the objective function values at optimality repor ted in Carolan et al. (1990), Meyer and Schultz (1991), Marsten et al. (1990), Setiono (1992) only agree to the first two digits of the values repor ted by the LQP algo- r i thm and MINOS. This led us to conclude that

Table 5 Performance of the LQP algorithm on a set of three PDS problems

Problem Major Simpl. No. of Subproblem Master Total iters iters vertices time time time

PDS1 7 22 16 7.5 104.18 111.68 PDS3 10 43 33 63.09 943.59 1006.68 PDS5 12 71 56 324.33 6641.33 6966.66

Page 15: A smooth penalty function algorithm for network-structured problems

234 S.A. Zenios et al. / A smooth penalty function algorithm for network structured

Table 6 Vector performance of the LQP algorithm on a set of three PDS problems

Test problem Scalar optimization Vectorization

Compiler User

PDS1 6.80 5.27 1.86 PDS3 83.33 44.68 18.77 PDS5 352.98 153.50 93.10

we have slightly different data from the rest of the literature.

Table 7 Solution times on the CRAY Y-MP

Test Simpl. No. of Subprob. Master Total Problem i ters vertices time time time

PDS1 23 15 1.01 0.85 1.86 PDS3 41 34 12.04 6.73 1&77 PDS5 71 54 55.12 37.97 93.09 PDS10 103 86 232.31 175.82 408.13 PDS15 121 104 559.35 381.08 940.43 PDS20 145 119 1225.83 720.06 1945.89 KENll 16 9 7.4 8.l 15.5 KEN13 87 14 66.8 70.6 137.5

4.6. Vector comput ing

It is well established in the scientific comput- ing literature that vector computer architectures can have a significant impact on the way algo- rithms perform. In this section we illustrate the performance of the LQP algorithm on a vector supercomputer CRAY Y-MPE264. Details on the vectorization of this algorithm - and additional research on a parallel computing implementation - are reported in Pmar and Zenios (1992). Our intention here is to illustrate the suitability of the algorithm for vector computations.

Table 6 illustrates the solution times on a CRAY Y-MP when executing in three distinct modes: 1) scalar optimization, 2) vector optimiza- tion with the CRAY CF77 compiler, and 3) vec- torization of the software by the designer of the algorithm as discussed in Pmar and Zenios (1992). We observed from this table that the algorithm achieves improvements in performance in the range 1.92-3.76 when implemented in vector mode as opposed to scalar.

point out that the interior point algorithms do no exploit the network structure, although they take advantage of the block diagonal sparsity pattern of the constraint matrix. The results of Marsten et al. and G E N O S / M C are directly comparable since they were both executed on the same com- puter and both codes were vectorized.

The same test problems were also solved by Carolan et al. (1990) on the KORBX system using several variants of Karmarkar's algorithm. The comparison between the KORBX system and G E N O S / M C does not discriminate between the relative merits of the algorithms and the underlying hardware platform. We feel, neverthe- less, that such a comparison is appropriate since both hardware systems are priced at approxi- mately the same range.

6. Conclusions

We have developed in this paper an algorithm for solving optimization problems with embedded network structures. Implementations for multi-

5. Solving large scale models

As a concluding exercise we solved all the test problems on the C R A Y Y-MP. The results are given in Table 7.

Some of these problems were solved on the same computer by Marsten et al. (1990) using the code OB1 based on interior point methods. The results are summarized in Table 8.

It is clear that the LQP algorithm outperforms substantially state-of-the-art implementations of interior point methods. It is, of course, fair to

Table 8 Comparative solution times of test problems with various methods in CPU hours (hrs: min: sec)

Test KORBX OB1 GENOS/MC problem

KENll 0:06:36 0:00:21 0:00:15

KEN13 0:20:45 0:01:07 0:02:17

PDSI0 3:18:00 0:25:30 0:06:48

PDS20 24:00:00 4:27:00 0:32:25

Page 16: A smooth penalty function algorithm for network-structured problems

S,4. Zenios et al. / A smooth penalty function algorithm for network structured 235

commodity network flow problems have been proven to be very effective and efficient in solving some large optimization problems. Since the method outperforms by a large margin state-of- the-art implementations of interior point meth- ods we can conclude that there is a high payoff in exploiting both macro-structures (i.e., block diag- onal) and micro-structures (i.e., network con- straints).

Acknowledgements

The assistance of Mr. J. Gregory with the experiments on the CRAY is gratefully acknowl- edged. Partial support was provided by NSF grants CCR-91-04042, and AFOSR grant 91,0168.

References

Ali, I., and Kennington, J.L. (1977), "MNETGN program documentation", Technical Report IEOR 77003, Depart- ment of Industrial Engineering and Operations Research, Southern Methodist University, Dallas, TX.

Assad, A.A. (1987), "Multicommodity network flows: A sur- vey", Networks 8, 37-92.

Bertsekas, D.P. (1975), "Nondifferentiable optimization via approximation", Mathematical Programming Study 3, 1-25.

Bertsekas, D.P. (1982), Constrained Optimization and La- grange Multiplier Methods, Academic Press, New York.

Bixby, R.E., and Fourer, R. (1988), "Finding embedded net- work rows in linear programs I. Extraction heuristics", Management Science 34, 342-376.

Carolan, W.J., Hill, J.E., Kennington, J.L., Niemi, S., and Wichmann, S.J. (1990), "An empirical evaluation of the KORBX algorithms for military airlift applications", Oper- ations Research 38, 240-248.

Dembo, R.S. (1987), "A primal truncated Newton algorithm with application to nonlinear network optimization", Mathematical Programming Study 31, 43-72.

Dembo, R.S., and Klincewicz, J.G. (1981), "A scaled reduced gradient algorithm for network flow problems with convex separable costs", Mathematical Programming Study 15, 125-147.

Dembo, R.S., and Klincewicz, J. (1985), "Dealing with degen- eracy in reduced gradient algorithms", Mathematical Pro- gramming 31, 357-363.

Dembo, R.S., Mulvey, J.M., and Zenios, S.A. (1989), "Large scale nonlinear network models and their application", Operations Research 37, 353-372.

Fiacco, A.V., and McCormick, G.P. (1968), Nonlinear Pro- gramming: Sequential Unconstrained Minimization Tech- niques, Wiley, New York.

Gamble, A.B., Conn, A.R., and Pulleyblank, W.R. (1991), "A network penalty problem", Mathematical Programming 50, 53-74.

Gill, P., Murray, W., and Wright, M.H. (1981), Practical Optimization, Academic Press, London.

Hearn, D.W., Lawphongpanich, S., and Ventura, J.A. (1987), "Restricted simplicial decomposition: Computation and extensions", Mathematical Programming Study 31, 99-118.

Holloway, C.A. (1974), "An extension of the Frank-Wolfe method of feasible directions", Mathematical Program- ming 6, 14-27,

Karmarkar, N. (1984), "A new polynomial time algorithm for linear programming", Combinatorica, 4, 373-395.

Kennington, J.L. (1978), "A survey of linear multicommodity network flows", Operations Research 26, 209-236.

Marsten, R., Subramanian, R., Saltzman, M., Lustig, I., and Shanno, D. (1990), "Interior point methods for linear programming", Interfaces 20, 105-116.

Meyer, R.R., and Zenios, S.A. (eds.) (1988), Parallel Opti- mization on Novel Computer Architectures, Annals of Oper- ations Research 14, J.C. Baltzer AG, Scientific Publishing Company, Basel, Switzerland.

Mulvey, J.M., and Vladimirou, H. (1991), "Solving multistage stochastic networks: An application of scenario aggrega- tion", Networks 21, 619-644.

Mulvey, J.M., and Zenios, S.A. (1987), "User's guide to GENOS 1.0: A Generalized Network Optimization Sys- tem", Report 87-12-03, Department of Decision Sciences, University of Pennsylvania, Philadelphia, PA.

Mulvey, J.M., Zenios, S.A., and Ahlfeld, D.P. (1990), a simpli- cial decomposition of convex generalized networks", Jour- nal of Information and Optimization Sciences 11, 359-387.

Murtagh, B.A., and Sannders, M.A. (1983), "MINOS 5.0 user's guide", Technical report SOL 83-20, Systems Opti- mization Laboratory, Stanford University, Stanford, CA.

Nielsen S.S., and Zenios, S.A. (1993), "A massively parallel algorithm for nonlinear stochastic network programs", Op- erations Research 41, 319-337.

Pmar, M.C., and Zenios, S.A. (1993), "Solving nonlinear programs with embedded network structure", in: D.-Z. Du and P. Pardalos (eds.), Network Models and Algorithms, World Scientific, Series on Applied Mathematics, Vol. 2, 177-202.

Pmar, M.C., and Zenios, S.A. (1994), "On smoothing exact penalty functions for convex constrained optimization", SIAM Journal on Optimization, to appear.

Plnar, M.C., and Zenios, S.A. (1992), "Parallel decomposition of multicommodity network flows using linear-quadratic penalty functions", ORSA Journal on Computing 4, 235- 249.

Ruszcynski, A. (1989), "An augmented lagrangian decomposi- tion method for block diagonal linear programming prob- lems", Operations Research Letters 8, 287-294.

Setiono, R. (1992), "Interior proximal point algorithm using preconditioned conjugate gradient", Optimization. Mathe- matische Operationsforsehung und Statistik 24, 63-73.

Schultz, G.L., and Meyer, R.R. (1991), "An interior point method for block-angular optimization", SIAM Journal on Optimization 1, 583-602.

Page 17: A smooth penalty function algorithm for network-structured problems

236 S.A. Zenios et al. / A smooth penalty function algorithm for network structured

Von Hohenbalken, B. (1977), "Simplicial decomposition in nonlinear programming algorithms", Mathematical Pro- gramming 13, 49-68.

Zang, I. (1981), "Discontinuous optimization by smoothing", Mathematics of Operations Research 6, 140-152.

Zenios, S.A., and Pmar, M.C. (1992), "Parallel block parti- tioning of truncated Newton for nonlinear network opti-

mization", SIAM Journal on Scientific and StatiStical Com- puting 13, 1173-1193.

Zenios, S.A., Qi, R., and Armstrong, P.K. (1990), "Coercion methods for decomposition of nonlinear programs: I. Basic theory and computations", Report 90-11-09, Department of Decision Sciences, The Wharton School, University of Pennsylvania, Philadelphia, PA.