global convergence and ascent property of a cyclic ... · i.a. geraldo, a. n’guessan and k. e....

13
Afrika Statistika Vol. 13 (2), 2018, pages 1631–1643. DOI: http://dx.doi.org/10.16929/as/1631.125 Afrika Statistika ISSN 2316-090X Global convergence and ascent property of a cyclic algorithm used for statistical analysis of crash data Issa Cherif Geraldo 1,* , Assi N’Guessan 2 and Kossi Essona Gneyou 3 1 Ecole d’ing´ enieurs en Informatique et Technologies num´ eriques, Universit´ e catholique de l’Afrique de l’Ouest - Unit´ e universitaire du Togo, 01 B.P. 1502 Lom´ e 01, Lom´ e, Togo. 2 Laboratoire Paul Painlev´ e UMR CNRS 8524, Universit´ e de Lille 1, 59655 Villeneuve d’Ascq Cedex, France. 3 epartement de Math´ ematiques, Facult´ e des Sciences, Universit´ e de Lom´ e, B.P. 1515 Lom´ e, Togo. Received Febrary on 02, 2018, Accepted on March 17, 2018 Copyright c 2018, Afrika Statistika and The Statistics and Probability African Soci- ety (SPAS). All rights reserved Abstract. In this paper, we consider an estimation algorithm called cyclic iterative algorithm (CA) that is used in statistics to estimate the unknown vector parameter of a crash data model. We provide a theoretical proof of the global convergence of the CA that justifies the good numerical results obtained in early numerical studies of this algorithm. We also prove that the CA is an ascent algorithm, what ensures its numerical stability. esum´ e. Dans cet article, nous consid´ erons un algorithme d’estimation appel´ e al- gorithme cyclic iteratif (CA) utilis´ e en statistique pour estimer le vecteur param` etre inconnu d’une mod` ele pour les donn´ ees d’accidents. Nous donnons une preuve th´ eorique de la convergence globale du CA qui justifie les excellents r´ esultats num´ eriques obtenues dans les ´ etudes num´ eriques ant´ erieures dudit algorithme. Nous prouvons aussi que le CA augmente la log-vraisemblance ` a chaque it´ eration, ce qui assure sa stabilit´ e num´ erique. Key words: Iterative method, Maximum likelihood, cyclic algorithm, global con- vergence, road safety. AMS 2010 Mathematics Subject Classification : 62-04, 62F10, 62H12, 62P99. * Corresponding Author: [email protected] [email protected] Kossi [email protected]

Upload: others

Post on 22-Nov-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

Afrika StatistikaVol. 13 (2), 2018, pages 1631–1643.DOI: http://dx.doi.org/10.16929/as/1631.125

Afrika Statistika

ISSN 2316-090XGlobal convergence and ascent property of acyclic algorithm used for statistical analysisof crash dataIssa Cherif Geraldo1,∗, Assi N’Guessan2 and Kossi Essona Gneyou3

1Ecole d’ingenieurs en Informatique et Technologies numeriques, Universite catholique del’Afrique de l’Ouest - Unite universitaire du Togo, 01 B.P. 1502 Lome 01, Lome, Togo.2Laboratoire Paul Painleve UMR CNRS 8524, Universite de Lille 1, 59655 Villeneuve d’AscqCedex, France.3Departement de Mathematiques, Faculte des Sciences, Universite de Lome, B.P. 1515Lome, Togo.

Received Febrary on 02, 2018, Accepted on March 17, 2018

Copyright c© 2018, Afrika Statistika and The Statistics and Probability African Soci-ety (SPAS). All rights reserved

Abstract. In this paper, we consider an estimation algorithm called cyclic iterativealgorithm (CA) that is used in statistics to estimate the unknown vector parameterof a crash data model. We provide a theoretical proof of the global convergence ofthe CA that justifies the good numerical results obtained in early numerical studiesof this algorithm. We also prove that the CA is an ascent algorithm, what ensuresits numerical stability.

Resume. Dans cet article, nous considerons un algorithme d’estimation appele al-gorithme cyclic iteratif (CA) utilise en statistique pour estimer le vecteur parametreinconnu d’une modele pour les donnees d’accidents. Nous donnons une preuvetheorique de la convergence globale du CA qui justifie les excellents resultatsnumeriques obtenues dans les etudes numeriques anterieures dudit algorithme.Nous prouvons aussi que le CA augmente la log-vraisemblance a chaque iteration,ce qui assure sa stabilite numerique.

Key words: Iterative method, Maximum likelihood, cyclic algorithm, global con-vergence, road safety.AMS 2010 Mathematics Subject Classification : 62-04, 62F10, 62H12, 62P99.

∗Corresponding Author: [email protected]@polytech-lille.frKossi [email protected]

Page 2: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1632

1. Introduction

In statistics, one is usually confronted to the problem of the estimation of the un-known parameter vector β ∈ Rd of a probability law Lβ. The first method and alsothe most used for this purpose is the Maximum Likelihood (ML) method (Cramer,1946; Shao, 2003). If Y1, . . . , Yn are n random variables considered as independentand identically distributed according to Lβ, the likelihood of observed data y1, . . . ,yn is the product of the probabilities that each Yi takes the value yi, i = 1, . . . , n.The maximum likelihood method consists in maximizing the likelihood functionL(β) considered as a function of β ∈ Rd. In practice, one prefers to maximize thelog-likelihood function ` = logL. Except some school examples, the maximum like-lihood estimation of β often requires the use of a numerical iterative optimizationmethod starting from an initial vector β(0) and computing successive approxima-tions of the unknown solution by the recurrence formula

β(k+1) = M(β(k)) (1)

where M is a mapping from Rd into itself. Detailed reviews of modern numericaloptimization methods can be found in Dennis and Schnabel (1996), Nocedal andWright (2006), Lange (2013) and Lange et al. (2014).

The very first method that comes to mind is the Newton-Raphson’s (NR) algo-rithm. The NR algorithm for the ML estimation of β uses the iteration mapping

M(β(k)) = −(∇2`(β(k))

)−1∇`(β(k)) (2)

where ∇` is the gradient of ` and ∇2` is its Hessian matrix. Since the NR algorithmrequires the inversion of the Hessian matrix at each iteration, its implementationmay be difficult in large dimensions or if the Hessian matrix is ill-conditioned orsingular at the point β(k). We can also mention the fact that the NR algorithm candiverge violently when the starting point β(0) is far from the true unknown valueof β (Dennis and Schnabel, 1996). When the NR algorithm fails, one can use oneof the many algorithms that have been proposed in the literature as remedies.One can use the Fisher scoring algorithm which replaces the Hessian matrix bythe expectation of its negative (Osborne, 1992) or quasi-Newton algorithms whichcompute an approximation of the inverse of the Hessian matrix using only thefirst derivatives (Nocedal and Wright, 2006). We can also mention Derivative FreeOptimization (DFO) in which no derivative of the objective function is computedand the successive iterates are computed from the values of the objective functionon a finite set of points (Rios and Sahinidis, 2013). DFO algorithms are of interestwhen ` is expensive to evaluate or non-differentiable. The recent decades have alsoseen the popularization of the Expectation Maximization (EM) algorithm (Demp-ster et al., 1977; McLachlan and Krishnan, 2008) which is considered as a specialcase of the more general class of MM (Minorization-Majorization) optimizationalgorithms (Hunter and Lange, 2004; Zhou and Lange, 2010). In maximizationproblems, the first M step consists in minorizing the objective function ` by a

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 3: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1633

surrogate function g(β|β(k)) and the second M step consists in maximizing this sur-rogate function with respect to β to produce the next iterate β(k+1). MM algorithmsare considered as effective algorithms for ML estimation because they consistentlydrive the likelihood uphill by maximizing a simple surrogate function for the log-likelihood. However, all these remedies brought by scientific results come at thecost of a greater and greater complexity and they are not always easy to implement.

Regardless of the algorithm used, global convergence is necessary. An optimizationalgorithm is said to be globally convergent if the sequence of iterates β(k) generatedby this algorithm converges to a stationary point of ` (a point where the gradientvanishes) from any starting β(0) (Dennis and Schnabel, 1996). Proving the globalconvergence is always a delicate exercise and this property may not always be sat-isfied. For example, the NR algorithm is not globally convergent because the con-vergence to a stationary point of the sequence β(k) generated by the NR algorithmis guaranteed only if the starting point β(0) is sufficiently close to the true value ofβ (Lange, 2013). For MM algorithms, the global convergence has been establishedunder some conditions (Lange, 2010). In order to ensure numerical stability, amaximum likelihood estimation algorithm should also be an ascent algorithm i.e.it should increase the log-likelihood at each step.In the context of statistical analysis of crash data, N’Guessan and Truffier (2008)have considered a parameter vector β = (θ, φT )T where θ is the first componentof β and φ is a vector consisting in the remaining components of β. They provedthe existence of a unique stationary point and proposed an estimation algorithmcalled Cyclic iterative Algorithm (CA) that cycles through the two subsets ofcomponents updating each from the other one. More specifically, this CA startsfrom an initial value φ(0) and updates successively θ(1) from φ(0), φ(1) from θ(1), θ(2)from φ(1) and so on until a convergence criteria is satisfied. N’Guessan and Geraldo(2015) studied some numerical convergence properties of the CA using simulatedaccident datasets. Their simulation studies suggest that the CA converges tothe maximum likelihood estimator (MLE) β of β from any starting point β(0) andoutperforms the classical optimization algorithms such as NR and MM. Geraldoet al. (2015) proved that the MLE β is strongly consistent, that is, it convergesalmost surely to the true value of β when the sample size tends to +∞.

In this paper, we provide convergence results that justify the good numericalresults given by N’Guessan and Geraldo (2015) and which complete the stochasticconvergence results of Geraldo et al. (2015). We prove that the CA is globallyconvergent (it converges to the MLE from any starting value β(0)) and is also anascent algorithm (the log-likelihood increases at each iteration of the algorithmi.e. `(β(k+1)) > `(β(k)) for any iteration k > 0).

The rest of the paper is structured as follows. Section 2 provides the state of theart of the cyclic algorithm. In section 3, we provide some intermediate lemmasthat will be used to prove the main convergence theorems given in Section 4. Moreprecisely, we prove that the sequence of iterates β(k) generated by the CA converges

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 4: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1634

to the MLE from every starting point and that the CA enjoys the ascent property.The paper finishes with some concluding remarks.

2. An overview of the cyclic algorithm

2.1. Problem setting and statistical model

Consider an experimental site where accidents can be classified into r (r > 0)mutually exclusive accidents types by increasing severity (for example, propertydamage, minor injury, severe injury, fatal). Assume that a road safety measure(transformation of intersections into roundabouts, installation of roundabouts,modification of the ground marking, etc. . . ) has been applied at this experimentalsite with the purpose of reducing the occurrence of accidents. After a certainperiod of application of the measure, it is certainly desirable to know if the latterhad a significant effect.

There exists in the literature a plethora of statistical models for the evaluationof a road safety measure. In general, these models strongly depend on the avail-able data. Lord and Mannering (2010) and Mannering and Bhat (2014) providecomprehensive reviews of contemporary thinking in the crash frequency-analysisfield and give the advantages and disadvantages of each approach. In this work,we consider the before-after model proposed by N’Guessan et al. (2001). Oneof the main advantages of this before-after model is that it allows cause-effectinterpretations (Hauer, 2010).

Let X = (X11, . . . , X1r, X21, . . . , X2r)T be a random vector where X1j (resp. X2j ),

j = 1, . . . , r, represents the number of crashes of type j occurred in the ”before”(resp. ”after”) period. In order to take into account some external factors (traf-fic flow, speed limit variation, weather conditions, etc. . .), the experimental siteis associated to a control area where the safety measure was not applied. LetZ = (z1, . . . , zr)

T be a vector such that zj denotes the ratio of the number of ac-cidents of type j for the period ”after” to the period ”before” in the control areaover the same time period. The model of N’Guessan et al. (2001) (in the case of oneexperimental site) is described by the following assumptions:

(A1) The total number of accidents observed on the experimental site where themeasure was applied is a fixed constant denoted n.

(A2) The control coefficients z1, . . . , zr are known and non-random.(A3) The random vector X is assumed to have the multinomial distribution

X ∼M(n;π1(β), π2(β))

where β = (θ, φT )T ∈ R1+r, θ > 0, φ = (φ1, . . . , φr)T belongs to the simplex

Sr ={

(φ1, . . . , φr)T ∈ Rr | φi > 0, 1 6 i 6 r,

r∑i=1

φi = 1}, (3)

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 5: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1635

πi(β) = (πi1(β), . . . , πir(β))T , i = 1, 2 and

πij(β) =

φj

1 + θ∑rm=1 zmφm

, i = 1; j = 1, . . . , r,

θzjφj1 + θ

∑rm=1 zmφm

, i = 2; j = 1, . . . , r

(4)

The parameter θ represents the mean effect of the road safety measure while thecomponents of φ = (φ1, . . . , φr)

T represent the different accident risks.

2.2. The cyclic algorithm (CA) for maximum likelihood estimation of the parameters

For an observed data x = (x11, . . . , x1r, x21, . . . , x2r) such that∑2i=1

∑rj=1 xij = n, set

x·j = x1j + x2j (j = 1, . . . , r) and xi· =∑rj=1 xij (i = 1, 2). Then the log-likelihood is

defined up to an additive constant by

`(β) =

r∑j=1

(x·j log(φj) + x2j log(θ)− x·j log(1 + θ

r∑m=1

zmφm)

). (5)

Provided that it exists, the Maximum Likelihood Estimator, β = (θ, φT )T , of β =(θ, φT )T is solution to the constrained optimization problem:

β = argmaxβ∈R1+r

`(β)

subject to∀j = 1, . . . , r, φj > 0, θ > 0 and

∑rj=1 φj = 1.

(6)

N’Guessan et al. (2001) proved that the MLE β is solution to the non-linear systemof equations

r∑j=1

(x2j −

x·r θ∑rm=1 zmφm

1 + θ∑rm=1 zmφm

)= 0

x·j −nφj(1 + θzj)

1 + θ∑rm=1 zmφm

= 0, j = 1, . . . , r.

(7)

N’Guessan (2010) proved that the non-linear system (7) accepts a solution β =(θ, φT )T such that

θ =x2·

x1·

1∑rj=1 zj φj

φj =1

1− 1

n

r∑m=1

θzmx·m

1 + θzm

× x·j

n(1 + θzj), j = 1, . . . , r.

(8)

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 6: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1636

He then proposed an iterative procedure alternating between updating θ holding φfixed and vice-versa. Because of the link between θ and φ, his procedure starts froman initial vector φ(0) = (φ

(0)1 , . . . , φ

(0)r )T such that

∑rj=1 φ

(0)j = 1. At the step k+1, θ(k+1)

is updated from φ(k) = (φ(k)1 , . . . , φ

(k)r )T , afterwards φ(k+1) = (φ

(k+1)1 , . . . , φ

(k+1)r )T

is updated from the θ(k+1), and so on until the convergence of the sequence ofvectors β(k) = (θ(k), φ(k)

T)T is achieved. This strategy yields the following iterative

algorithm (N’Guessan, 2010):

Given an initial vector φ(0) = (φ(0)1 , . . . , φ

(0)r )T such that

∑rj=1 φ

(0)j = 1 and for any

k > 0,

θ(k+1) =x2·

x1·

1∑rj=1 zjφ

(k)j

φ(k+1)j =

1

1− 1

n

r∑m=1

θ(k+1)zmx·m1 + θ(k+1)zm

× x·jn(1 + θ(k+1)zj)

, j = 1, . . . , r.

(9)

The aim of this paper is to prove that the sequence of vectors β(k) = (θ(k), φ(k)T

)T

produced by Algorithm (9) converges to the MLE β = (θ, φT )T of β. We also provethat Algorithm (9) is an ascent algorithm i.e. for all k > 0, `(β(k+1)) > `(β(k)).

3. Preliminary results

Lemma 1. Let ψ be the mapping defined on R+ by

ψ(u) = −x1· +r∑

m=1

x·m1 + uzm

. (10)

i) There exists a unique real number θ∗ such that ψ(θ∗) = 0 and the MLE θ of θ isequal to that unique root θ∗ of ψ.

ii) For all u > 0, ψ(u) > 0 if 0 < u 6 θ∗ and ψ(u) 6 0 if u > θ∗.

Proof.i) One can easily check that ψ is continuous and its derivative ψ′(u) is strictlynegative for every u > 0 and therefore ψ is bijective. Moreover,

limu→0

ψ(u)× limu→+∞

ψ(u) = (x2·)× (−x1·) < 0.

Hence the equation ψ(u) = 0 has a unique solution denoted θ∗. Let j be an integerfrom the set {1, . . . , r}. From the equalities

φj =1

1− 1

n

r∑m=1

θzmx·m

1 + θzm

× x·j

n(1 + θzj)

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 7: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1637

and

1− 1

n

r∑m=1

θzmx·m

1 + θzm= 1− 1

n

r∑m=1

(x·m −

x·m

1 + θzm

)=

1

n

r∑m=1

x·m

1 + θzm,

we may write

φj =x·j/(1 + θzj)∑r

m=1 x·m/(1 + θzm)(11)

and thenr∑j=1

zj φj =

∑rj=1

(zjx·j/(1 + θzj)

)∑rm=1

(x·m/(1 + θzm)

) .From the first line of (8), we deduce

θ =x2·x1·

∑rm=1

(x·m/(1 + θzm)

)∑rj=1

(zjx·j/(1 + θzj)

) .This is equivalent to

x2·x1·

r∑m=1

x·m

1 + θzm=

r∑m=1

θzmx·m

1 + θzm.

As we have x1· + x2· = n, we deduce the following equality:

n

x1·

r∑m=1

x·m

1 + θzm= n

which yields the equality ψ(θ) = 0. Since ψ is bijective and ψ(θ∗) = 0 then θ = θ∗.

ii) The function ψ is a strictly decreasing function. Therefore,

∀u 6 θ∗, ψ(u) > ψ(θ∗) = 0 and ∀u > θ∗, ψ(u) 6 ψ(θ∗) = 0.

This completes the proof of Lemma 1.

Lemma 2. Let αx = x2·/x1· and Ψx be the function from ]0; +∞[ to ]0; +∞[ defined by:

Ψx(u) = αx

(r∑

m=1

x·m1 + uzm

)/( r∑m=1

zmx·m1 + uzm

).

i) The real θ∗ defined by Lemma 1 is the unique fixed point of Ψx.ii) The function Ψx is such that

∀u 6 θ∗, Ψx(u) > u and ∀u > θ∗, Ψx(u) 6 u.

iii) The sequence of real numbers (θ(k)) generated by Algorithm (9) is monotonousand its monotony depends on θ(0). If θ(0) < θ∗ then it is an increasing sequenceand if θ(0) > θ∗, it is a decreasing sequence.

iv) The sequence (θ(k)) is also bounded. Then it is convergent and its limit is θ∗.

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 8: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1638

Proof.i) The equation Ψx(u) = u is equivalent to

x2·x1·

r∑m=1

x·m1 + uzm

=

r∑m=1

uzmx·m1 + uzm

.

After straightforward computations similar to those used in the proof of Lemma1, one gets ψ(u) = 0. This latter equation has a unique solution θ∗ that is also theunique solution of the equation Ψx(u) = u.

ii) Let u > 0. Then

Ψx(u)− u =

(x2·x1·

r∑m=1

x·m1 + uzm

−r∑

m=1

uzmx·m1 + uzm

)/( r∑m=1

zmx·m1 + uzm

)

=

(x2·x1·

r∑m=1

x·m1 + uzm

− n+

r∑m=1

x·m1 + uzm

)/( r∑m=1

zmx·m1 + uzm

)

=n

x1·

(r∑

m=1

x·m1 + uzm

− x1·

)/( r∑m=1

zmx·m1 + uzm

)

=n

x1·(ψ(u))

/( r∑m=1

zmx·m1 + uzm

).

The sign of Ψx(u)− u is obtained from that of ψ(u).

iii) From the second line of Equation (9) and the equality

1− 1

n

r∑m=1

θ(k+1)zmx·m1 + θ(k+1)zm

= 1− 1

n

r∑m=1

(x·m −

x·m1 + θ(k+1)zm

)=

1

n

r∑m=1

x·m1 + θ(k+1)zm

,

we deduce that the real sequence θ(k) is given by :

θ(k+1) =

(r∑

m=1

x·m1 + θ(k)zm

)/( r∑m=1

zmx·m1 + θ(k)zm

). (12)

which is equivalent to θ(k+1) = Ψx(θ(k)). The function Ψx is differentiable and itsderivative has the form Ψ′x(u) = αx ×Ψ1,x(u)/Ψ2,x(u) where

Ψ2,x(u) =

(r∑

m=1

zmx·m1 + uzm

)2

> 0

and

Ψ1,x(u) = −r∑i=1

zix·i(1 + uzi)2

r∑j=1

zjx·j1 + uzj

+

r∑i=1

x·i1 + uzi

r∑j=1

(zj)2x·j

(1 + uzj)2.

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 9: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1639

By swapping indexes i and j in the second right-hand term, we get

Ψ1,x(u) =

r∑i=1

r∑j=1

(zi)2x·ix·j − zizjx·ix·j

(1 + uzi)2(1 + uzj).

After removing the zero terms corresponding to i = j, we get:

Ψ1,x(u) =∑

16i<j6r

(zi)2x·ix·j − zizjx·ix·j

(1 + uzi)2(1 + uzj)+

∑16j<i6r

(zi)2x·ix·j − zizjx·ix·j

(1 + uzi)2(1 + uzj).

By swapping indexes i and j in the second right-hand term, we get

Ψ1,x(u) =∑

16i<j6r

(zi)2x·ix·j − zizjx·ix·j

(1 + uzi)2(1 + uzj)+

∑16i<j6r

(zj)2x·ix·j − zizjx·ix·j

(1 + uzj)2(1 + uzi).

It is easy to check that

Ψ1,x(u) =∑

16i<j6r

((zi)

2x·ix·j − zizjx·ix·j(1 + uzi)2(1 + uzj)

+(zj)

2x·ix·j − zizjx·ix·j(1 + uzi)(1 + uzj)2

)

=∑

16i<j6r

(x·ix·j

(1 + uzi)(1 + uzj)

)((zi)

2 − zizj1 + uzi

+(zj)

2 − zizj1 + uzj

).

Since

(zi)2 − zizj

1 + uzi+

(zj)2 − zizj

1 + uzj=

(zi)2 − 2zizj + (zj)

2

(1 + uzi)(1 + uzj)=

(zi − zj)2

(1 + uzi)(1 + uzj)> 0,

we get Ψ1,x(u) > 0. Thus the function Ψx is an increasing function.If θ(0) < θ∗, then by using the property ii) of Lemma 2, we will have

θ(0) < Ψx(θ(0)) = θ(1). As Ψx is an increasing function, we will then haveθ(1) = Ψx(θ(0)) < Ψx(θ(1)) = θ(2), θ(2) = Ψx(θ(1)) < Ψx(θ(2)) = θ(3) and so on. By asimilar reasoning, one can prove by recurrence that if θ(0) > θ∗ then θ(k) > θ(k+1)

for all k = 0, 1, 2, . . ..

iv) For every k > 0, we have

0 < θ(k) 6 max

(θ(0), sup

u>0Ψx(u)

)where

supu>0

Ψx(u) = limu→+∞

Ψx(u) =αxn

r∑m=1

x·mzm

. (13)

The real sequence (θ(k)) is monotonous and bounded. Thus it converges to θ∗ theonly fixed point of the function Ψx that is also equal to the MLE θ (by Lemma 1).The proof of Lemma 2 is then completed.

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 10: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1640

Because of the partition of the parameter vector β into two sub-parameters θ and φ,we consider the concentrated (or profile) likelihood function that is also commonlyused in maximum likelihood estimation (Monahan, 2011). For a given value of θ,the MLE of the sub-parameter φ is found as a function φ = g(θ) = (g1(θ), . . . , gr(θ))

T

where

gj(θ) =1

1− 1

n

r∑m=1

θzmx·m

1 + θzm

× x·j

n(1 + θzj), j = 1, . . . , r

=x·j/(1 + θzj)∑r

m=1 x·m/(1 + θzm), j = 1, . . . , r (by expression (11)).

(14)

The likelihood function `(β) = `(θ, φ) can be re-written as a function only of θ,

`c(θ) = `(θ, g(θ))

that is called the concentrated (or profile) likelihood.

Lemma 3. The concentrated (or profile) likelihood function is defined up to an addi-tive constant by

`c(θ) = x2· log θ −r∑j=1

x·j log(1 + θzj) (15)

Proof. The expression (5) is equivalent to

`(β) =

r∑j=1

x·j log(φj) + x2· log(θ)− n log(1 + θ

r∑m=1

zmφm)

and one can write

`c(θ) =r∑j=1

x·j log

(x·j

1 + θzj

)−

r∑j=1

x·j log

(r∑

m=1

x·m1 + θzm

)+ x2· log(θ)

− n log

(r∑

m=1

x·m1 + θzm

+

r∑m=1

θzmx·m1 + θzm

)+ n log

(r∑

m=1

x·m1 + θzm

).

This is equivalent to

`c(θ) =

r∑j=1

x·j log

(x·j

1 + θzj

)− n log

(r∑

m=1

x·m1 + θzm

)+ x2· log(θ)

− n log n+ n log

(r∑

m=1

x·m

1 + θzm

).

After removing the second and the fifth terms and the constants, we get the equality(15).

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 11: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1641

4. Main results

In this section we prove that the cyclic algorithm (9) satisfies two main propertiesgenerally required for ML estimation iterative algorithms. The first one is the con-vergence of the iterative scheme (9) to the MLE β = (θ, φT )T from the algorithmicviewpoint. Indeed, there is no guarantee that an iterative algorithm will convergeto the MLE. It should be noted that the convergence studied here is different fromthe consistency that was already studied by Geraldo et al. (2015). The first mainresult proved in this section is stated by the following theorem.

Theorem 1. For all starting vector β(0) = (θ(0), (φ(0))T )T where θ(0) > 0 and φ(0) ∈ Sr,the sequence β(k) = (θ(k), (φ(k))T )T generated by the cyclic algorithm (9) converges tothe MLE β = (θ, φT )T of β.

Proof. From Lemma 2, we can conclude that the sequence (θ(k)) converges to θ.Since each φ

(k)j , j = 1, . . . , r, is the image of θ(k) by the continuous mapping gj

defined by (14), the real sequence φ(k)j also has a limit that is gj(θ) = φj. Thus, thevector β(k) = (θ(k), (φ(k))T )T converges to the MLE β = (θ, φT )T . Theorem 1 is thusproved.

The second result that we prove is the ascent property of the cyclic algorithm (9)i.e. the fact that the log-likelihood is increased monotonically by the algorithm.That is given by the following theorem.

Theorem 2. The cyclic algorithm (9) enjoys the ascent property i.e. the sequenceβ(k) = (θ(k), (φ(k))T )T generated by the cyclic algorithm satisfies the property

`(β(k+1)) > `(β(k)), k = 0, 1, . . . (16)

Proof. The profile log-likelihood `c(θ) is differentiable for every θ > 0 and its deriva-tive is

`′c(θ) =x2·θ−

r∑j=1

x·jzj1 + θzj

=1

θ

x2· − r∑j=1

(x·j −

x·j1 + θzj

) =1

θ

x2· − n+r∑j=1

x·j1 + θzj

.

Since x1· + x2· = n, we have`′c(θ) = θ−1 ψ(θ)

where the function ψ is defined by equation (10) (Lemma 1). Using this lemma, wededuce that

∀θ 6 θ∗, `′c(θ) > 0 and ∀θ > θ∗, `′c(θ) 6 0

where θ∗ is the MLE of θ and also the unique root of ψ. Hence the function `c isincreasing on the interval ]0, θ∗] and decreasing on [θ∗,+∞[. To finish the proof, weconsider the two cases θ(0) < θ∗ and θ(0) > θ∗.

− If θ(0) < θ∗, then we have proved that the sequence θ(k) is increasing and stillbelongs to the interval ]0, θ∗]. Then θ(k) 6 θ(k+1) and `c(θ

(k)) 6 `c(θ(k+1)) because

`c is increasing on ]0, θ∗].

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 12: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1642

− If θ(0) > θ∗, then the sequence θ(k) is decreasing and still belongs to the interval[θ∗,+∞[. Then θ(k+1) 6 θ(k) and `c(θ

(k+1)) > `c(θ(k)) because `c is decreasing on

[θ∗,+∞[.

In all the cases, we have

`(β(k+1))− `(β(k)) = `c(θ(k+1))− `c(θ(k)) > 0

and the proof of Theorem 2 is complete.

5. Concluding remarks

In this paper, we gave some theoretical convergence theorems for a cyclic algo-rithm (CA) for the maximum likelihood estimation of the vector parameter of a sta-tistical model for crash data. The vector parameter is partitioned under the formβ = (θ, φT )T where θ is a positive real number while φ is a vector belonging to amultidimensional simplex. We proved the global convergence of the CA (i.e. theCA converges to the MLE of β from every initial point) and we also proved that itenjoys the ascent property. Future research will include the study of the conver-gence rate as well as the extension of our results to the statistical model proposedin N’Guessan et al. (2001) which is a generalization of the model studied in thispaper.Acknowledgment The author wish to thank the unknown reviewer for the in-sightful comments and recommendations that helped to improve the quality andthe editors for a high professional handling of the publication process.

References

Cramer, H. (1946). Mathematical methods of statistics. Princeton University Press,Princeton.

Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood fromincomplete data via the EM algorithm. Journal of the Royal Statistical Society,Series B, 39(1):1–38.

Dennis, Jr, J. E. and Schnabel, R. B. (1996). Numerical Methods for UnconstrainedOptimization and Nonlinear Equations. Classics in Applied Mathematics. Societyfor Industrial and Applied Mathematics.

Geraldo, I. C., N’Guessan, A., and Gneyou, K. E. (2015). A note on the strongconsistency of a constrained maximum likelihood estimator used in crash datamodelling. Comptes Rendus Mathematique, 353(12):1147–1152.

Hauer, E. (2010). Cause, effect and regression in road safety: A case study. AccidentAnalysis and Prevention, 42:1128–1135.

Hunter, D. R. and Lange, K. (2004). A Tutorial on MM Algorithms. The AmericanStatistician, 58(1):30–37.

Lange, K. (2010). Numerical Analysis for Statisticians. Springer.Lange, K. (2013). Optimization. Springer, New York, 2nd edition.Lange, K., Chi, E. C., and Zhou, H. (2014). A Brief Survey of Modern Optimization

for Statisticians. International Statistical Review, 82(1):46–70.

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst

Page 13: Global convergence and ascent property of a cyclic ... · I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages 1631 – 1643. Global convergence

I.A. Geraldo, A. N’Guessan and K. E. Gneyou, Afrika Statistika, Vol. 13 (2), 2018, pages1631 – 1643. Global convergence and ascent property of a cyclic algorithm used forstatistical analysis of crash data. 1643

Lord, D. and Mannering, F. (2010). The statistical analysis of crash-frequency data:A review and assessment of methodological alternatives. Transportation ResearchPart A, 44:291–305.

Mannering, F. L. and Bhat, C. R. (2014). Analytic methods in accident research:Methodological frontier and future directions. Analytic Methods in Accident Re-search, 1:1–22.

McLachlan, G. and Krishnan, T. (2008). The EM Algorithm and Extensions. Wi-ley series in probability and statistics. John Wiley & Sons, Inc., Hoboken, NewJersey, 2nd edition.

Monahan, J. F. (2011). Numerical Methods of Statistics. Cambridge UniversityPress, 2nd edition.

N’Guessan, A. (2010). Analytical Existence of solutions to a system of non-linearequations with application. Journal of Computational and Applied Mathematics,234:297–304.

N’Guessan, A., Essai, A., and Langrand, C. (2001). Estimation multidimensionnelledes controles et de l’effet moyen d’une mesure de securite routiere. Revue destatistique appliquee, 49(2):85–102.

N’Guessan, A. and Geraldo, I. C. (2015). A cyclic algorithm for maximum likelihoodestimation using Schur complement. Numerical Linear Algebra with Applications,22(6):1161–1179.

N’Guessan, A. and Truffier, M. (2008). Impact d’un amenagement de securiteroutiere sur la gravite des accidents de la route. Journal de la Societe Francaisede Statistique, 149(3):23–41.

Nocedal, J. and Wright, S. J. (2006). Numerical optimization. Springer, secondedition.

Osborne, M. R. (1992). Fisher’s method of scoring. International Statistical Review/ Revue Internationale de Statistique, 60(1):99–117.

Rios, L. M. and Sahinidis, N. V. (2013). Derivative-free optimization: a review ofalgorithms and comparison of software implementations. Journal of Global Op-timization, 56(3):1247–1293.

Shao, J. (2003). Mathematical Statistics. Springer, second edition.Zhou, H. and Lange, K. (2010). MM algorithms for some discrete multivariate

distributions. Journal of Computational and Graphical Statistics, 19(3):645–665.

Journal home page: www.jafristat.net, www.projecteuclid.org/as,www.ajol.info/afst