limit points of independent copies of sample maxima

7
Statistics & Probability Letters 65 (2003) 103 – 109 Limit points of independent copies of sample maxima George Mathew Department of Mathematics, Southwest Missouri State University, 901 South National Avenue, Springeld, MO 65804-0094, USA Received February 2003; received in revised form March 2003 Abstract Let {X (i) n ;n ¿ 1}, i =1; 2 be two independent copies of a sequence of independent and identically distributed random variables. Let M i n = max 16j6n X (i) j . For suitable normalizing constants a n and b n , we investigate the set of almost sure limit points of the vector sequence Z n = M 1 n b n a n ; M 2 n b n a n ; n ¿ 1 and further investigate the limit set for vector subsequences Z n k . c 2003 Elsevier B.V. All rights reserved. Keywords: Sample maxima; Set of limit points; Independent copies 1. Introduction Let {X (i) n ;n ¿ 1}, i =1; 2 be two independent copies of a sequence of independent and identically distributed random variables. Let M i n = max 16j6n X (i) j . Hebbar (1980) has investigated the set of almost sure limit points of vector sequences with independent components based on the maxima of a stationary Gaussian sequence. He showed that the limit set was a triangular region thus providing for the maxima another example of the phenomenon that when considering limit points of vector sequences with independent components, the resulting limit set is not the product set of the set of limit points for each component. For sums, the asymptotic behavior is discussed in a quite general setting in LePage (1973). Nayak (1985) addresses for general distributions the same problem as Hebbar but in the i.i.d. setting and takes as his starting point the one-dimensional result the law of the iterated logarithms provided by De Haan and Hordijk (1972). Also, another look at the results of De Haan and Hordijk E-mail address: [email protected] (G. Mathew). 0167-7152/$ - see front matter c 2003 Elsevier B.V. All rights reserved. doi:10.1016/j.spl.2003.07.006

Upload: george-mathew

Post on 26-Jun-2016

213 views

Category:

Documents


0 download

TRANSCRIPT

Statistics & Probability Letters 65 (2003) 103–109

Limit points of independent copies of sample maximaGeorge Mathew

Department of Mathematics, Southwest Missouri State University, 901 South National Avenue,Spring�eld, MO 65804-0094, USA

Received February 2003; received in revised form March 2003

Abstract

Let {X (i)n ; n¿ 1}, i=1; 2 be two independent copies of a sequence of independent and identically distributed

random variables. Let Min = max16j6n X

(i)j . For suitable normalizing constants an and bn, we investigate the

set of almost sure limit points of the vector sequence

Zn =(M 1

n − bnan

;M 2

n − bnan

); n¿ 1

and further investigate the limit set for vector subsequences Znk .c© 2003 Elsevier B.V. All rights reserved.

Keywords: Sample maxima; Set of limit points; Independent copies

1. Introduction

Let {X (i)n ; n¿ 1}, i=1; 2 be two independent copies of a sequence of independent and identically

distributed random variables. Let Min = max16j6n X

(i)j . Hebbar (1980) has investigated the set of

almost sure limit points of vector sequences with independent components based on the maxima ofa stationary Gaussian sequence. He showed that the limit set was a triangular region thus providingfor the maxima another example of the phenomenon that when considering limit points of vectorsequences with independent components, the resulting limit set is not the product set of the set oflimit points for each component. For sums, the asymptotic behavior is discussed in a quite generalsetting in LePage (1973).Nayak (1985) addresses for general distributions the same problem as Hebbar but in the i.i.d.

setting and takes as his starting point the one-dimensional result the law of the iterated logarithmsprovided by De Haan and Hordijk (1972). Also, another look at the results of De Haan and Hordijk

E-mail address: [email protected] (G. Mathew).

0167-7152/$ - see front matter c© 2003 Elsevier B.V. All rights reserved.doi:10.1016/j.spl.2003.07.006

104 G. Mathew / Statistics & Probability Letters 65 (2003) 103–109

was taken by HAusler (1985). His point of view was to examine the rate of maximal growth of themaxima when sampled along a subsequence nk . More precisely, he elucidated the eBect of takingsubsequences on the magnitude of Climk→∞M ∗

nk where M ∗n is the maximum suitably normalized.

In this paper, we consider a problem similar to the one considered by HAusler but directed towardsvector subsequences. We show that if the subsequence {nk ; k¿ 1} satisEes limk→∞ (nk−1)=nk ¿ 0then the entire set of almost sure limit points of

{Znk ; k¿ 1}={(

M 1nk − bnkank

);(M 2

nk − bnkank

); k¿ 1

}(1.1)

coincides with the limit set for the whole sequence. However, for thin subsequences, that is, forsubsequences satisfying limk→∞ (nk−1)=nk =0, we show that the upper boundary of the limit set for{Znk ; k¿ 1} is contained in the interior of the limit set for {Z n; n¿ 1}.We present the results in the next section.

2. Results

Let F denote the common distribution function of the sequence of random variables {X (i)n ; n¿ 1}.

Set

f(x) =1− F(x)F ′(x)

; (2.1)

where it is assumed that F ′(x) exists and is positive for all suFciently large x, and further set

g(x) = f(x) log log[1− F(x)]−1: (2.2)

From the work of De Haan and Hordijk (1972), the asymptotic behavior of g(x) as x → ∞ playsan important role in determining sample path properties of max16j6n Xj. Of considerable use is thefollowing identity giving the rate of decrease of the tail of a distribution belonging to the class ofdistributions considered in this paper. Let

bn = F−1(1− 1

n

)and an = f(bn) log log x = g(bn); (2.3)

where f is deEned in (2.1). Suppose in (2.2) that g′ exists and

limt→∞ g′(t) = 0:

Then, by De Haan and Hordijk (1972, p. 1192), we have

1− F(bn + xan) =(log n)rn

n; (2.4)

where limn→∞ rn(x) = −x. In order to prove our results we need the following result (Pathak andQualls, 1973, Lemma 2.5).

Lemma 2.1 (Pathak and Qualls). Let {In; n¿ 1} be a sequence of indicator functions such that∑∞n=1 EIn =∞. If for some A, 06A¡∞, Var(∑n

1 Ik)6A(∑n

1 EIk)2 for all n su7ciently large,

then P(∑∞

1 Ik =∞)¿ 1=(1 + A).

G. Mathew / Statistics & Probability Letters 65 (2003) 103–109 105

To reduce the notational burden, we let Un = (M 1n − bn)=an and Vn = (M 2

n − bn)=an, where an andbn are deEned in (2.3).

Theorem 2.2. Suppose F is twice di8erentiable with F ′(x) positive for all real x, and let r ¡ 1. Iflimt→∞ g′(t) = 0, then we have∑

k

(log nk)−r ¡∞ implies P{Unk ¿x; Vnk ¿y; i:o:}= 0 (2.5)

for x and y such that x + y¿r, and∑k

(log nk)−r =∞ implies P{Unk ¿x; Vnk ¿y; i:o:}= 1 (2.6)

for x and y such that x + y¡r.

Proof. Observe that for r ¡x + y we have

P{Unk ¿x; Vnk ¿y} = [1− Fnk (bnk + xank )][1− Fnk (bnk + yank )]

= (1 + o(1))(log nk)rnk (x)+rnk (y)

6 (log nk)−r for all k suFciently large; (2.7)

where we have used relation (2.4). Thus by (2.7), it is clear that (2.5) holds by the Borel–CantelliLemma.Suppose the condition in (2.6) holds where r ¡ 1 and x+ y¡r, for some x; y. Clearly, to show

P{Unk ¿x; Vnk ¿y; i:o:}=1, it suFces to establish P{Unk ¿x; Vnk ¿y; i:o:}=1 for any subsequence{tk ; k¿ 1}. In particular, consider the subsequence deEned by

ntk = [exp(kr−1)]: (2.8)

Let Ik = I{Untk¿ x; Vntk

¿y}. Then as in the Erst part of the proof, we have for all suFcientlylarge k

EIk = P{Untk¿ x; Vntk

¿y}¿ (log ntk )

−�r for some 0¡�¡ 1 such that x + y¡�r

¿ k−�; (2.9)

where the last inequality follows from (2.8). Therefore by (2.9), it follows that∑k

EIk =∞: (2.10)

Next observe that as noted by De Haan and Hordijk (1972, p. 1192), bn + xan is ultimately non-decreasing in n so that it follows that for j¡k suFciently large, and with n′j = ntj

P{Un′j ¿ x; Vn′j ¿y;Un′k ¿ x; Vn′k ¿y}= p1p2; (2.11)

106 G. Mathew / Statistics & Probability Letters 65 (2003) 103–109

where

p1 =P{M 1n′j¿ bn′k + xan′k}

+P{(bn′j + xan′j ¡M 1n′j6 bn′k + xan′k ) ∩ (M 1

n′j ;n′k¿bn′k + xan′k )} (2.12)

with Mn′j ;n′k= maxn′j¡i6n′k Xi. Also, p2 is similarly deEned with M 2

n in place of M1n and y in place

of x.Letting un = un(x) = bn + xan, we can write (2.12) as

p1 = 1− Fn′j(un′j)− Fn′k (un′k ) + Fn′j(un′j)Fn′k−n′j(un′k ): (2.13)

Similarly, we have an expression for p2 with un(y) in place of un(x).Also, in the same way one Ends that

P{Un′j ¿ x; Vn′j ¿y}P{Un′k ¿ x; Vn′k ¿y}= q1q2; (2.14)

where

q1 = [1− Fn′j(un′j)][1− Fn′k (un′k )] (2.15)

and where again q2 is likewise deEned with un(y) in place of un(x).Let di = pi − qi, i = 1; 2 and observe by deEnitions (2.13) and (2.15)

d1 = Fn′j(un′j)[Fn′k−n′j(un′k )− Fn′k (un′k )] (2.16)

with a similar expression for d2.Then by (2.11) and (2.14), we have

Cov(Ij; Ik) = p1p2 − q1q2 = q1d2 + q2d1 + d1d2: (2.17)

Further, it can be checked that for suFciently large j; k

q16 (log n′j)−x=2(log n′k)

−x=2 (2.18)

and

d16 (n′j=n′k)(log n

′k)

−x=2 (2.19)

with similar expressions for q2 and d2. Therefore in view of the bounds in (2.18) and (2.19), itfollows that Cov(Ij; Ik) is at most 3(n′j=n′k).Now, observe that

Var

(m∑1

Ij

)=

m∑j=1

Var(Ij) + 2m−1∑j=1

Cov(Ij; Ij+1) + 2m−2∑j=1

m∑k=j+2

Cov(Ij; Ik)

6 5m∑1

EIj + 6m−2∑j=1

m∑k=j+2

(n′j=n′k): (2.20)

The above inequality holds since for indicator random variables Ij, Var(Ij)6EIj,∑m−1

j=1 Cov(Ij; Ij+1)6 2

∑mj=1 EIj, and by the above stated upper bound for Cov(Ij; Ik). Let

’k = exp(kr−1): (2.21)

G. Mathew / Statistics & Probability Letters 65 (2003) 103–109 107

Let sm = [m%], where 0¡%¡ 1− �. Then,m−2∑j=1

m∑k=j+2

(n′j=n′k)6

m−2∑j=1

m−1∑k=j+1

(’j=’k)

6sm∑j=1

m∑k=j+1

(’j=’k) +m−1∑

j=sm+1

m∑k=j+1

(’j=’k)

6 smm∑

k=1

(’1=’k+1) + mm∑

k=1

(’sm=’sm+k); (2.22)

where the last inequality follows since

max{(’j=’k): k − j = t and j¿ j0}= (’j0=’j0+t):

Now as∑∞

1 ’1=’k+1¡∞ and ’sm=’sm+k ¡ exp{−r−1sr−1−1m k}, the Enal expression in (2.22) isbounded above by

csm + mm∑

k=1

exp{−r−1sr−1−1m k}6 c′

m∑1

EIk ; (2.23)

where c=∑m

k=1 (’1=’k+1) and c′ a suitable constant. The last inequality follows from the deEnitionof sm and (2.9).Therefore by (2.10), (2.22), and (2.23) we have that for any '¿ 0 and m suFciently large

Var

(m∑1

Ij

)6 '

(m∑1

EIj

)2:

Therefore by Lemma 2.1, it follows that P{Untk¿ x; Vntk

¿y; i:o:}= 1, which completes the proofof Theorem 2.2.

Now we present the result for subsequences.

Theorem 2.3. If the subsequence {nk ; k¿ 1} satis�eslimk→∞

(nk−1=nk)¿ 0; (2.24)

then the set of almost sure limit points of {(Unk ; Vnk ); k¿ 1},lim set (Unk ; Vnk ) = {(x; y): 06 x; y and x + y6 1}: (2.25)

On the other hand if the subsequence {nk ; k¿ 1} satis�es the conditionslimk→∞

(nk−1=nk) = 0 (2.26)

∑k

(log nk)a ¡∞ if a¡− r and∑k

(log nk)a =∞ if a¿− r (2.27)

108 G. Mathew / Statistics & Probability Letters 65 (2003) 103–109

for some r ¡ 1, then the set of points {(x; r − x): 06 x6 r} forms the upper boundary of thelimit set; that is, for any x′; y′ such that x′ + y′¿r we have

P{Unk ¿x′; Vnk ¿y′; i:o:}= 0; (2.28)

and for any x′; y′ such that x′ + y′¡r

P{Unk ¿x′; Vnk ¿y′; i:o:}= 1: (2.29)

Proof. Assume condition (2.24) holds and suppose for some 0¡x; y that r = x + y¡ 1. To showthat (x; y) is a limit point of the sequence (Unk ; Vnk ) it suFces to establish for some subsequence(n′k ; k¿ 1)

P{Un′k ¿ x′; Vn′k ¿y′; i:o:}= 0 for any x′; y′ with x′ + y′¿x + y (2.30)

and

P{Un′k ¿ x′; Vn′k ¿y′; i:o:}= 1 for any x′; y′ with x′ + y′¡x + y: (2.31)

We choose as our subsequence the subsequence deEned in (2.8). Observe that under assumption(2.24) it follows that for all k suFciently large, ntk ¡ntk+1 so that for any r′¿r∑

k

(log ntk+1)−r′ 6

∑k

(log’k)−r′

=∑k

k−(r′=r)

¡∞:

Therefore, by Theorem 2.2 we have (2.30) holds. Similarly we End for any r′¡r that∑k

(log ntk )−r′ ¿

∑k

(log’k)−r′

=∑k

k−(r′=r)

=∞:

Hence, an appeal to Theorem 2.2 yields (2.31).Thus we have shown that

lim set (Unk ; Vnk ) ⊃ {(x; y): 0¡x; y and x + y¡ 1}:Therefore, since the limit set for the entire sequence is {(x; y): 06 x; y and x+ y6 1} as given inTheorem 3.2 of Nayak (1985), it follows that (2.25) holds.Finally, since condition (2.27) implying (2.28) and (2.29) is just a restatement of Theorem 2.2,

the proof of Theorem 2.3 is complete.

Remark. Let {X (i)n ; n¿ 1}, i=1; 2; : : : ; ‘ be ‘ independent copies of a sequence of independent and

identically distributed random variables, and for ‘¿ 2, let {Z nk ; k¿ 1} be deEned similarly as in(1.1). Then, an extension of Theorem 2.3 to the case of ‘¿ 2 can be obtained just by making the

G. Mathew / Statistics & Probability Letters 65 (2003) 103–109 109

following two changes: the limit set in (2.25) becomes {(x1; x2; : : : ; x‘): 06 xi and∑‘

1 xi6 1}; andif (2.27) holds, then the upper boundary becomes {(x1; x2; : : : ; x‘): 06 xi and

∑‘1 xi = r}.

Acknowledgements

The author is grateful to Dr. W.P. McCormick for valuable comments and suggestions.

References

De Haan, L., Hordijk, A., 1972. The rate of growth of sample maxima. Ann. Math. Statist. 43, 1185–1196.Hebbar, H.V., 1980. Almost sure limit points of maxima of stationary Gaussian sequences. Ann. Probab. 8, 393–399.HAusler, J., 1985. The growth rate of partial maxima on subsequences. Statist. Probab. Lett. 3, 245–249.LePage, R.D., 1973. Log log law for Gaussian processes. Z. Wahrsch. Verw. Gebiete 25, 103–108.Nayak, S.S., 1985. Almost sure limit points of independent copies of sample maxima. Stochastic Process. Appl. 20,353–360.

Pathak, P.K., Qualls, C., 1973. A law of iterated logarithm for stationary Gaussian processes. Trans. Amer. Math. Soc.181, 185–193.