• No results found

A Berry Esseen type bound for the kernel density estimator based on a weakly dependent and randomly left truncated data

N/A
N/A
Protected

Academic year: 2020

Share "A Berry Esseen type bound for the kernel density estimator based on a weakly dependent and randomly left truncated data"

Copied!
19
0
0

Loading.... (view fulltext now)

Full text

(1)

R E S E A R C H

Open Access

A Berry-Esseen type bound for the kernel

density estimator based on a weakly

dependent and randomly left truncated data

Petros Asghari and Vahid Fakoor

*

*Correspondence:

fakoor@math.um.ac.ir

Department of Statistics, Faculty of Mathematical Sciences, Ferdowsi University of Mashhad, P.O. Box 1159-91775, Mashhad, Iran

Abstract

In many applications, the available data come from a sampling scheme that causes loss of information in terms of left truncation. In some cases, in addition to left truncation, the data are weakly dependent. In this paper we are interested in deriving the asymptotic normality as well as a Berry-Esseen type bound for the kernel density estimator of left truncated and weakly dependent data.

Keywords: left-truncation; weakly dependent; asymptotic normality; Berry-Esseen;

α

-mixing

1 Introduction

P is a population with large, deterministic and finite sizeN with elements{(Yi,Ti);i= , . . . ,N}. In sampling from this population we only observe those pairs for whichYiTi. Suppose that there is at least one pair with this condition. The sample is denoted by

{(Yi,Ti);i= , . . . ,n}. This model is called random left-truncated model (RLTM). We as-sume that{Yi;i≥}is a stationaryα-mixing sequence of random variables and{Ti;i= , . . . ,N}is an independent and identically distributed (i.i.d.) sequence of random vari-ables. The definition of a strong mixing sequence is presented in Definition .

Definition  Let{Yi;i≥}be a sequence of random variables. The mixing coefficient of this sequence is

α(m) =sup

k≥

P(AB) – P(A)P(B);AFk,BFk+m,

whereFlmdenotes theσ-algebra generated by{Yj}forljm. This sequence is said to be strong mixing orα-mixing if the mixing coefficient converges to zero asm→ ∞.

Studying the various aspects of left-truncated data is of high interest due to their ap-plicability in much research. One of these applications is in survival analysis. It is well known that in medical research on some specific diseases such as AIDS and dementia, the sampling scheme results in data samples that are left truncated. This model also arises in astronomy [].

(2)

Strong mixing sequences of random variables are widely occurring in practice. One application is in the analysis of time series and in renewal theory. A stationary ARMA-sequence fulfils the strong mixing condition with an exponential rate of mixing coefficient. The concept of strong mixing sequences was first introduced by Rosenblatt [] where a central limit theorem is presented for a sequence of random variables that satisfies the mixing condition.

The Berry-Esseen inequality or theorem was stated independently by Berry [] and Es-seen []. This theorem specifies the rate at which the scaled mean of a random sample converges to the normal distribution for all sample spaces. Parzen [] derived a Berry-Esseen inequality for the kernel density estimator of an i.i.d. sequence of random variables. Several works were done for left-truncated observations. We can refer to [] where the dis-tribution of left-truncated data was estimated and asymptotic properties of the estimator were derived. More work was done by Stute []. Prakasa Rao [] presented a Berry-Esseen theorem for the density estimator of a sample that forms a stationary Markov process. Liang and U¨na-Álvarez [] have derived a Berry-Esseen inequality for mixing data that are right censored. Yang and Hu [] presented Berry-Esseen type bounds for kernel den-sity estimator based on aϕ-mixing sequence of random variables. Asghariet al.[, ] presented a Berry-Esseen type inequality for the kernel density estimator, respectively, for a left-truncated model and for length-biased data.

This paper is organized as follows. In Section , needed notations are introduced and some preliminaries are listed. In Section , the Berry-Esseen type theorem for the esti-mator of the density function of the data is presented. In Section , the theorems and corollaries of Section  are proved.

2 Preliminaries and notation

Suppose thatYi’s andTi’s fori= , . . . ,Nare positive random variables with distributions FandG, respectively. Let the joint distribution function of (Y,T) be

H∗(y,t) = P(Y≤y,T≤t)

= 

α

y

–∞G(tu)dF(u),

in whichα= P(Y≥T).

If the marginal distribution function ofYiis denoted byF∗, we have

F∗(y) = 

α

y

–∞

G(u)dF(u),

so the marginal density function ofYis

f∗(y) = 

αG(y)f(y).

A kernel estimator forf is given by

fn(y) =  nhn

n

i=

K

Yiy hn

α

(3)

In many applications, the distribution function of the truncation random variableGis unknown. Sofn(y) is not applicable in these cases and we need to use an estimator ofG. Before starting the estimation details, for any distribution functionLon [,∞], letaL:=

inf{x>  :L(x) > }andbL:=sup{x>  :L(x) < }.

Woodroof [] pointed out thatFandGcan be estimated only ifaGaF,bGbF and ∞

aF

dF

G <∞. This integrability condition can be replaced by the stronger conditionaG<aF. Using this assumption, here we use the non-parametric maximum likelihood estimator for Gthat is presented by Lynden-Bell [] and is denoted byGn,

Gn(y) = i:Yi>y

 – S(y) Cn(y)

, ≤y<∞, (.)

in whichS(y) =ni=I{Yi=y}andCn(s) = 

n

n

i=I{TisYi}.

Using the definition ofCnthat is mentioned in the estimation procedure ofGand also using the empirical estimators ofF∗andG∗, which are denoted byFn∗andGn, we have

Cn(y) =Gn(y) –Fn∗(y), y∈[aF, +∞).

It can be seen that Cn(s) is actually the empirical estimator of Cn =G∗(y) –F∗(y) =

α–G(y)[ –F(y)],y[aF, +). This fact gives the following estimator ofα:

αn=

Gn(y)[ –Fn(y)] Cn(y) .

For details as regardsαn, see []. Usingαn, we present a more applicable estimator off, which is denotedfnˆ and is defined as

ˆ

fn(y) =  nhn

n

i=

K

Yiy hn

αn

Gn(Yi). (.)

Note that in (.) the sum is taken overi’s for whichGn(Yi)= .

3 Results

Before presenting the main theorems, we need to state some assumptions. Suppose that aG<aF andbGbF. Woodroof [] stated that the uniform convergence rate ofGntoG is true fory∈[a,bG] fora>aG. Thus, we have to assume thataG<aF. LetC= [a,b] be a compact set such thatC⊂ {y;y∈[aF,bF[}. As mentioned in the Introduction,{Yi;i≥} is a stationaryα-mixing sequence of random variables with mixing coefficientβ(n), and

{Ti;i≥}is an i.i.d. sequence of random variables.

Definition  The kernel functionK, is a second order kernel function if K(t)dt= ,

–∞tK(t)dt=  and –∞∞tK(t)dt> .

Assumptions

A β(n) =O(n–λ)for someλ>+δ

δ in which <δ≤.

A For the conditional density ofYj+givenY=y(denoted byfj(·|y)), we havefj(y|y)≤

(4)

A (i) Kis a positive bounded kernel function such thatK(t) = for|t|> and

–K(t) = .

(ii) Kis a second order kernel function. (iii) f is twice continuously differentiable.

A Letp=pnandq=qnbe positive integers such thatp+qn, there exists a constantC

such that fornlarge enough qpC. Alsophn→,qhn→asn→ ∞.

A {Ti;i≥}is a sequence of i.i.d. random variable with common continuous distribution

functionG, and independent of{Yi;i≥}.

H The kernel functionK(·)is differentiable and Hölder continuous with exponentβ> . H β(n) =O(n–λ)forλ>+ββ in whichβ>.

H The joint density of(Yi,Yj),fij∗, exists and we havesupu,v|fij∗(u,v) –f∗(u)f∗(v)| ≤C<∞

for some constantC.

H There existsλ> +β and for the bandwidthhnwe havelog logn

nhn →andCn

(–λ)β β(λ+)+β++η<

hn<Cn–λ whichηis such that 

β(λ+)+β+<η< (λ–)β β(λ+)+β++

 –λ.

Discussion of the assumptions. A, A, and A are common in the literature. For exam-ple Zhou and Liang [] used A for deconvolution estimator of multivariate density of

α-mixing process. A(i)-(ii) are commonly used in non-parametric estimation. A(iii) is specially needed for a Taylor expansion. H-H are needed to use Theorem . of [] in Theorem  here.

Letσn(y) :=nhnVar[fn(y)], so by letting √ nhnK(

Yiy

hn ) α

G(Yi)=:Wni, we can write

σn(y) =Var

n

i=

nhnK

Yiy hn

α

G(Yi)

=Var

n

i=

Wni

. (.)

Letk= [pn+q],km= (m– )(p+q)+ andlm= (m– )(p+q)+p+, in whichm= , , . . . ,k. Now we have the following decomposition:

n

i=

Wni=Jn+Jn+Jn, (.)

in which

J n=

k

m=

jnm, jnm= km+p–

i=km

Wni,

J n =

k

m=

jnm, jnm= lm+q–

i=lm

Wni,

J

n =jnk+, jnk+=

n

i=k(p+q)+

Wni.

From now on, we letσ(y) :=αf(y)

G(y) 

–K(t)dt,u(n) :=

j=n(α(j))

(5)

Theorem  If AssumptionsA-A(i)andAare satisfied and f and G are continuous in a neighborhood of y for yaF,then for large enough n we have

sup

x∈R

Pnhnfn(y) –Efn(y)≤xσn(y)

– (x)=O(an) a.s.

in which

an=h (+δ)

+δ

n

p nhn

+δ

+λnα(q)/+λn/+λn/+h–δ/(+δ)n u(q), (.)

and

λn:=kq n +h

–δ/(+δ)

n u(q) +qhn,

λn :=p

n(phn+ ).

(.)

Theorem  If the assumptions of TheoremandAare satisfied,then for yaF and for large enough n we have

sup

x∈R

Pnhnfnˆ(y) –Efn(y)≤xσn(y)

– (x)=Oan+ (hnlog logn)/ a.s.

in which anis defined in(.).

Theorem  If the assumptions of Theoremare satisfied,G has bounded first derivative

in a neighborhood of y and f has bounded derivative of orderin a neighborhood of y for yaF,then for large enough n we have

sup

x∈R

Pnhnfnˆ(y) –f(y)≤(y)– (x)=Oan a.s.,

in which

an:=h (+δ)

+δ

n

p nhn

+δ

+hn(p+ ) +hδ

+δ

n u(q) +

λnα(q)/

+λn/+λn/+λn/λn/, (.)

andλandλare defined in(.).

Remark  In many applications,f andGare unknown and should be estimated, soσ(y) is not applicable in these cases. Here we present an estimator for it that is denoted byσˆn(y) and is defined as follows:

ˆ

σn(y) = αnfnˆ(y) Gn(y)

–

K(t)dt.

Using this estimator instead ofσ(y) in Theorem , costs a change in the rate of

(6)

Corollary  Let AssumptionsA, AandH-Hbe satisfied,then for yC

sup

yaF

σˆn(y) –σ(y)=O(cn) a.s.,

in which

cn:=max

logn nhn,h

n

+

log logn

n . (.)

Theorem  Let AssumptionsA-AandH-Hbe satisfied.For yCand for large enough n we have

sup

x∈R

Pnhn ˆ

fn(y) –f(y)

ˆn(y)

– (x)=Oan+cn

a.s.,

in which anis defined in(.)and cnis defined in(.).

4 Proofs

In order to start the proofs of the main theorems, we shall state some lemmas that are used in the proving procedure of the main theorems. For the sake of simplicity letC,CandC, be positive appropriate constants which may take different values at different places.

Lemma ([]) Let X and Y be random variables such that E|X|r<∞and E|Y|s<∞in which r and s are constants such that r,s> and r–+s–< .Then we have

E(XY) –E(X)E(Y)≤ XrYs

sup

Aσ(X),Bσ(Y)

P(AB) – P(A)P(B)–/r–/s.

Lemma  Suppose that AssumptionsA-A(i)andAare satisfied.If f and G are con-tinuous in a neighborhood of y for yaF thenσn(y)→σ(y)as n→ ∞.Furthermore,if f and G have bounded first derivatives in a neighborhood of y for yaF,for such y’s we have

σn(y) –σ(y)=O(bn),

in which

bn:=hn(p+ ) +h–δ/(+δ)n u(q) +λn/+λn/+λn/λn/, Proof Using the decomposition that is defined in (.) we can write

σn(y) =VarJn+Jn+Jn

=VarJn+VarJn+VarJn

+ CovJn,Jn+ CovJn,Jn+ CovJn,Jn, (.)

VarJn=Var

k

m=

jnm

= k

m=

km+p–

i=km

(7)

+ 

≤i<jk

Covjni,jnj+  k

m=

kmi<jkm+p–

Cov(Wni,Wnj)

=: I+ II+ III. (.)

As assumed in the lemma,f andGare continuous in a neighborhood ofyso they are

bounded in this neighborhood. Now under Assumption A(i) we have

Var(Wni) =  nhn E KYiy

hn

α

G(Yi)

E

K

Yiy

hn

α

G(Yi) =  nhn K

uy hn

αf(u) G(u) du

K

uy

hn

f(u)du

= 

n

K(t)αf(y+thn) G(y+thn) dthn

K(t)f(y+thn)dt

, (.)

so it can be concluded that

I≤kp n

–

K(t)αf(y+thn) G(y+thn)

dt+hn

–

K(t)f(y+thn)dt

 =O kp n . (.)

Lemma  for arbitrarilyδ>  and also the continuity off in a neighborhood ofygives

II≤

≤i<jk ki+p–

s=ki

kj+p–

t=kj

Cov(Wns,Wnt)

C

nhn

≤i<jk ki+p–

s=ki

kj+p–

t=kj

K

Ysy

hn

α

G(Ys) +δ K Yty

hn

α

G(Yt)

+δ

×α(ts)–/(+δ)

= C

nhn

≤i<jk ki+p–

s=ki

kj+p–

t=kj

K

Y–y

hn

α

G(Y)

+δ

α(ts)δ/(+δ)

= C

nhn

K

Y–y

hn

α

G(Y)

+δ

≤i<jk ki+p–

s=ki

kj+p–

t=kj

α(ts)δ/(+δ),

now using the notationu(n) :=∞j=n(α(j))δ+δ , which is defined before, and A we get the following result:

II≤ C

nhn

K

Y–y

hn

α

G(Y)

+δ

≤i<jk ki+p–

s=ki

kj+p–

t=kj

α(ts)δ/(+δ)

Ckp

nhδ/(+δ)n

–

K(t)+δ f(y+thn) G+δ(y+th

n) dt

δ/(+δ)

j=q

α(j)δ/(+δ)

(8)

Under Assumption A we can write

III≤ k

m=

kmi<jkm+p–

Cov(Wni,Wnj)

≤ 

nhn k

m=

kmi<jkm+p–

EK

Yiy

hn

K

Yjy hn

α

G(Yi)G(Yj)

+E

K

Yiy hn

α

G(Yi)

≤ 

nhn k

m=

kmi<jkm+p–

K

u–y

hn

K

u–y

hn

×f∗(u|u)f∗(u)dudu+

K

uy

hn

f(u)du

Chn

n k

m=

kmi<jkm+p–

–

–

K(s)K(t)ds dt+

+

–

K(t)f(y+thn)dt

Ckp

hn

n

–

–

K(s)K(t)ds dt+

–

K(t)f(y+thn)dt

=O(phn). (.)

Now, using (.), (.), (.), and (.), we have

VarJn=O

kp n +h

–δ/+δ

n u(q) +phn

=O(), (.)

VarJn=Var

k

m=

jnm

= k

m=

lm+q–

i=lm

Var(Wni) + 

≤i<jk

Covjni,jnj

+  k

m=

lmi<jlm+q–

Cov(Wni,Wnj)

=: I+ II+ III. (.)

By the same argument as is used for|I|and|II|and|III|, it can be concluded that

I=O

kq n

,

II=Oh–δ/(+δ)n u(q), III=O(qhn).

(9)

Now, using (.) and (.), we have

VarJn=O

kq n +h

–δ/(+δ)

n u(q) +qhn

=n. (.)

Similarly

VarJn=Var

n

i=k(p+q)+

Wni

= n

i=k(p+q)+

Var(Wni) + 

k(p+q)+≤i<jn

Cov(Wni,Wnj)

=: I+ II, (.)

and

I=O

n

nk(p+q),

II=O

ph

n n

.

(.)

So we can write

VarJn=O

n

nk(p+q)+phn

=O

p

n(phn+ )

=n. (.)

Gathering all that is obtained above,

σn(y) –σ(y)=

n

i=

Var(Wni) –σ(y)

+ 

≤i<jk

Covjni,jnj+  k

m=

kmi<jkm+p–

Cov(Wni,Wnj)

+ 

≤i<jk

Covjni,jnj+  k

m=

lmi<jlm+p–

Cov(Wni,Wnj)

+ 

k(p+q)+≤i<jn

Cov(Wni,Wnj) + CovJn,Jn

+ CovJn,Jn+ CovJn,Jn

(10)

and by letting

An:=

≤i<jk

Covjni,jnj+ k

m=

kmi<jkm+p–

Cov(Wni,Wnj)

+

≤i<jk

Covjni,jnj+ k

m=

lmi<jlm+p–

Cov(Wni,Wnj)

+

k(p+q)+≤i<jn

Cov(Wni,Wnj) +CovJn,Jn

+CovJn,Jn+CovJn,Jn,

we have

(.)≤

n

i=

Var(Wni) –σ(y)

+ |An|. (.)

On the other hand using (.), (.), and (.), we have

CovJn,JnVarJnVarJn

=n/, CovJn,Jn=n/, CovJn,Jn=n/λn/.

(.)

So forAnwe can write

|An|=O

h–δ/+δn u(q) +phn+qhn+phn n +λ

/

n +λn/+λn/λn/

=Oh–δ/+δn u(q) +phn+qhn+λn/+λn/+λn/λn/. (.) On the other hand from (.), it can easily be concluded thatni=Var(Wni)→σ(y) as n→ ∞. Now under Assumptions A and A|An| →, soσ

n(y)→σ(y). Iff andGhave bounded first derivatives in a neighborhood ofy, we can write

n

i=

Var(Wni) –σ(y)

= hn

E

K

Y–y

hn

α

G(Y )

E

K

Y–y

hn

α

G(Y)

σ(y)

= hn

K

uy

hn

αf(u) G(u) du

K

uy

hn

f(u)du

σ(y)

=

–

K(t)G(y)[f(y+thn) –f(y)] +f(y)[G(y+thn) –G(y)] G(y+hnt)G(y)

dt

hn

–

K(t)f(y+thn)dt

(11)

From (.) we get the following result:

σn(y) –σ(y)

=Ohn+h–δ/+δn u(q) +phn+qhn+λn/+λn/+λn/λn/

=Ohn+h–δ/+δn u(q) +phn+λn/+λn/+λn/λn/

=Ohn(p+ ) +h–δ/+δn u(q) +λn/+λn/+λn/λn/

, (.)

and the proof is completed.

Before starting the next lemma, we note that

nhn[fn(y) –Efn(y)]

σn(y)

= 

σn(y)

nhn n

i=

K

Yiy

hn

α

G(Yi)–E

K

Yiy hn

α

G(Yi)

=: n

i=

Zni. (.)

If we letni=Zni=:Sn, it can be observed that

Sn=Sn+Sn+Sn, (.)

in which

Sn= k

m=

Ynm , Ynm = km+p–

i=km

Zni,

Sn= k

m=

Ynm , Ynm = lm+q–

i=lm

Zni,

S= k

m=

Ynm, Ynm = n

i=k(p+q)+

Zni.

Lemma  Suppose that AssumptionsA-A(i)andAare satisfied and f and G are con-tinuous in a neighborhood of y for yaF.Then for such y’s we have

PSn>λ

  n

=

  n

,

PSn>λ

  n

=

  n

.

Proof With the aid of Lemma  we can write

ESn= 

σ

n(y)

EJnEJn

(12)

The same argument shows thatE(Sn)=O(λ

n), so we have

PSn>λ

  n

E(Sn)

λ

  n

=

  n

(.)

and

PSn>λ

  n

E(Sn)

λ

  n

=

  n

. (.)

So the proof is completed.

In the following letHn:=km=Xnm in whichXnm,m= , . . . ,k, are independent ran-dom variables with the same distribution as Ynm , m= , . . . ,k. ϕ and ϕ are, respec-tively, the characteristic functions ofSnandHn. Also lets

n :=

k

m=Var(Xnm) andsn:=

k

m=Var(Ynm ).

Lemma  Under the assumptions of Lemma,for yaF we have the following:

sn– =n/+λn/+h–δ/(+δ)n u(q).

Proof It can easily be seen thats

n=E(Sn)– 

≤i<jkCov(Yni,Ynj),E(Sn) =  and

ESn– =ESn–ES

n

=ESn–ESn+Sn+Sn

=ESn+Sn– ESnSn+Sn. (.)

Using (.) and Lemma , we can write

sn– =ESn– 

≤i<jk

CovYni,Ynj– 

ESn– + 

≤i<jk

CovYni,Ynj

=ESn+Sn– ESnSn+Sn+ 

≤i<jk

CovYni,Ynj

=n/+λn/+ 

≤i<jk

Covjni,jnj. (.)

On the other hand, from Lemma  we know thati<jkCov(jni,jnj) =O(h–δ/(+δ)n u(q)), so substituting this in (.), gives the result,

(13)

Lemma ([]) Let{Xj,j≥}be a stationary sequence with mixing coefficientα(k)and suppose that E(Xn) = ,r> ,and there existτ > andλ>r(rτ+τ) such thatα(n) =O(n–λ)

and also E|Xi|r+τ <∞.In this case,for any> ,there exists a constant C,for which we have E n i= Xi rC n i=

E|Xi|r+

n

i= Xir

r/

.

Lemma  Under the assumptions of Lemmafor yaF we have

sup

x∈R P

Hn

snx

– (x)=O

h (+δ)

+δ n p nhn +δ .

Proof Using [], Theorem ., forr>  we can write

sup

x∈R P

Hn

snx

– (x)≤C

k

m=E|Xnm|r

sr n

. (.)

On the other hand, using Lemma  there existsτ>  such that for any>  k

m=

E|Xnm|r = k

m=

E|Ynm|r

= k m= E

km+p–

i=km

Zni rk m= C km+p–

i=km

E|Zni|r+ km+p–

i=km

Znir r

. (.)

Let=δ,r=  + δfor  < δ<δandτ=δ– δandλ>(+δδ–δ)(+δ) , so we have

(.) =C k

m=

p+δ

(nhn)+δ K

uy hn

(+δ)

f(u) G+δ(u)du

+ p

+δ

(nhn)+δ K

uy

hn

+δ

f(u) G+δ(u)du

(+δ) (+δ)

=C k m= p+δ

n+δhδ

n

–

K(t)+δ f(y+hnt) G+δ(y+hnt)dt

+ p

+δ

n+δh

δ(+δ) +δ

n

–

K(t)+δ f(y+hnt) G+δ(y+hnt)dt

(+δ) (+δ)

Ckp+δ(nhn)–(+δ

)

hn+p+δ

(nhn)–(+δ

)

h (+δ)

+δ n =O h

(+δ) +δ n p nhn +δ . (.)

(14)

Lemma ([]) Let{Xj,j≥}be a stationary sequence with mixing coefficientα(k). Sup-pose that p and q are positive integers.Let Tl=j(l=(–)(l–)(p+pq+)+q)+p Xjin which≤lk.If s,r>  such that s–+r–= ,there exists a constant C> such that

Eexp it k l= Tlk l=

Eexp(itTl)

C|t|α/s(q) k

l= Tlr.

Lemma  Under the assumptions of Lemmafor yaFwe have

sup

x∈R

PSnx– P(Hnx)=O

λnα(q)/+h (+δ)

+δ n p nhn +δ .

Proof By lettingb=  in [], Theorem ., p., for anyT>  we have

sup

x∈R

PSnx– P(Hnx)≤

T

T

ϕ(t) –tϕ(t)dt

+Tsup

x∈R

|u|≤TC

P(Hnu+x) – P(Hnx)du

=:Ln+Ln. (.)

Now by lettings=r=  in Lemma , there exists a constantC>  for which we have

ϕ(t) –ϕ(t)= Eexp it k m= Ynmk m=

Eexp(itYnm)

C|t|α(q)   k m= Ynm

=Ck|t|α(q)E

km+p–

i=km

Zni  , (.)

E(Zn)=

α

nhnσ

n(y) E

K

Y–y

hn

G(Y)

E

K

Y–y

hn

G(Y)

≤ 

n(y)

–

K(t)f(y+thn) G(y+thn)dt+hn

–

K(t)f(y+thn)dt

=On–, (.)

E(ZnZn)≤

nhnσn(y)

 –  – K y–u

hn

K

y–u

hn

f∗(y|y)f∗(y)dydy

=O hn n . (.)

Now using (.) and (.) we have

E

km+p–

i=km

Zni  = km+p–

i=km

E(Zni)+  kmi<jkm+p–

EZniZnj

=O

p

n( +phn)

(15)

so

Ln=O

T

(q)

n ( +phn)

/

=OTλnα(q)/. (.)

On the other hand applying Lemma  gives

sup

x∈R

P(Hnu+x) – P(Hnx)

≤sup

x∈R P

Hn

sn

u+x

sn

u+x sn

+sup

x∈R P

Hn

snx sn

x sn

+sup

x∈R

usn+x

x sn

=O

h

(+δ) +δ

n

p nhn

+δ

+O

| u| sn

, (.)

so

Ln=O

h

(+δ) +δ

n

p nhn

+δ

+ 

T

. (.)

By choosingT= (α(q)λn)–/we get the following result: sup

x∈R

PSnx– P(Hnx)

=O

h

(+δ) +δ

n

p nhn

+δ

+λnα(q)/

=O(bn), (.)

and the lemma is proved.

Lemma ([]) Let X and Y be random variables.For any a> we have

sup

t∈R

P(X+Yt) – (t)≤sup

t∈R

P(Xt) – (t)+√aπ + P

|Y|>a.

Proof of Theorem Using (.) and Lemma , for anya>  anda>  we can write

sup

x∈R

Pnhnfn(y) –Efn(y)≤xσn(y)

– (x)

=sup

x∈R

PSn+Sn+Snx– (x)

≤sup

x∈R

PSnx– (x)+√a π +

a √

π + PS

n>a

+ PSn>a

. (.)

By choosinga=λn

/

anda=λn/and using Lemma , we have (.) =sup

x∈R P

(16)

On the other hand using Lemmas , , and  we have

sup

x∈R

PSnx– (x)

≤sup

x∈R

PSnx– P(Hnx)+sup

x∈R

P(Hnx) – (x)

≤sup

x∈R

PSnx– P(Hnx)+sup

x∈R

P(Hnx) –

x sn

+sup

x∈R

sxn– (x)

=O

h

(+δ) +δ

n

p nhn

+δ

+λnα(q)/

+O

h

(+δ) +δ

n

p nhn

+δ

+Osn– 

=O

h

(+δ) +δ

n

p nhn

+δ

+λnα(q)/+λn/+λn/+h–δ/(+δ)n u(q)

. (.)

So the proof is completed.

Proof of Theorem According to Lemma  for anya>  we can write

sup

x∈R

P(nhn)/fnˆ(y) –Efn(y)≤xσn(y)

– (x)

≤sup

x∈R P

(nhn)/fn(y) –Efn(y)≤xσn(y)

– (x)

+√aπ+ P

nhnfn(y) –fn(y)|

σn(y)

>a

, (.)

and

P

nhnfn(y) –fn(y)|

σn(y)

>a

≤ 

aE

nhnfn(y) –fn(y)|

σn(y)

, (.)

E

nhnfn(y) –fn(y)|

σn(y)

≤√ 

nhnσn(y) n

i=

EK

Yiy hn

αn Gn(Yi)–K

Yiy

hn

α

G(Yi)

=√ 

nhnσn(y) n

i=

E

K

Yiy hn

αnG(Yi) –αGn(Yi) Gn(Yi)G(Yi)

≤√ 

nhnσn(y) n

i=

E

K

Yiy hn

G(Y)|αnα|+α|Gn(Y) –G(Y)|

Gn(Yi)G(Yi)

. (.)

From Lemma . of [] we have

|αnα|=O

log logn n

(17)

and from [] we have

sup

yaF

Gn(y) –G(y)=O

log logn n

a.s. (.)

So we can write

(.)≤Cnhn

log logn n

–

K(t)f(y+thn)dt

=O(hnlog logn).

Now by choosinga= (hnlog logn)/and using Theorem  we get the result

sup

x∈R

P(nhn)/ ˆ

fn(y) –E

fn(y)

xσn(y)

– (x)

=Oan+ (hnlog logn)/

. (.)

Proof of Theorem By the triangular inequality and using Lemma  for

a=

nhn|Efn(y) –f(y)|

σ(y) , we have

sup

x∈R

Pnhnfnˆ(y) –f(y)≤(y)– (x)

≤sup

x∈R

Pnhn

ˆ

fn(y) –f(y)

σn(y)

σ(y)

σn(y) x

σ(y)

σn(y) x

+sup

x∈R

σσn((yy))x

– (x)+

nhn|Efn(y) –f(y)|

π σ(y) . (.)

Here we used the fact that the event

nhn

σ(y) |Efn(y) –f(y)|>adoes not happen for the

se-lecteda.

From the inequalitysupy| (ηy) – (y)| ≤ 

e√π(|η– |+|η

–– |), it can be concluded

that

sup

x∈R

σσn((yy))x

– (x)=n(y) –σ(y). (.)

Under Assumptions A(ii) and A(iii), use of the Taylor expansion yields

Efn(y) –f(y)=Ohn. (.)

So from (.), (.), (.), Theorem , and Lemma , we have

sup

x∈R

Pnhnfnˆ(y) –f(y)≤(y)– (x)=Oan+bn+hn

(18)

Proof of Corollary Using the triangular inequality it can be seen that

sup

y∈C

σˆn(y) –σ(y)

=sup

y∈C αnfnˆ(y)

Gn(y)

αf(y) G(y)

–K(t)dt

≤sup

y∈C

G(y)|αnfnˆ(y) –αf(y)|+αf(y)|Gn(y) –G(y)| Gn(y)G(y)

–

K(t)dt. (.)

Under Assumptions A, A, H-H, Theorem . of [] we obtain

sup

y∈C

fnˆ(y) –f(y)=O

max

logn nhn

,hn

a.s. (.)

From (.) and (.) we have

sup

y∈C

αnfnˆ(y) –αf(y)=sup y∈C

αnfnˆ(y) –αnf(y) +αnf(y) –αf(y)

≤sup

y∈C

αnfnˆ(y) –f(y)+sup y∈C

f(y)|αnα|

=O

max

logn nhn

,hn

+O

log logn n

a.s. (.)

Using (.) and (.) in (.) proves the corollary.

Proof of Theorem Using the triangular inequality we can write

sup

x∈R P

nhn(fnˆ(y) –f(y))

ˆ

σn(y) ≤ x

– (x)

≤sup

x∈R P

nhn(fˆn(y) –Efn(y))

σ(y) ≤

ˆ

σn(y)

σ(y)x

ˆ

σn(y)

σ(y)x

+sup

x∈R

σˆn(y)

σ(y)x

– (x). (.)

By Assumptions A-A(i), A and A, Theorem  results in the following:

sup

x∈R P

nhn(fˆn(y) –Efn(y))

σ(y) ≤

ˆ

σn(y)

σ(y)x

ˆ

σn(y)

σ(y)x

=Oan, (.)

in whichanis defined in Theorem .

Under Assumptions A, A, H-H, Corollary  results in the following:

sup

x∈R

σˆn(y)

σ(y)x

– (x)=ˆn(y) –σ(y)

=O

max

logn nhn

,hn

+

log logn n

a.s. (.)

(19)

5 Conclusions

In this paper we obtained Berry-Esseen type bounds for the kernel density estimator based on left-truncated and strongly mixing data. Here it is concluded that in RLTM, which is also dealing with weak dependency, we can get asymptotic normality but comparing the results with [] we see that the rates get much more complicated and also slower.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors read and approved the final manuscript.

Acknowledgements

The authors would like to sincerely thank the anonymous referees for their careful reading of the manuscript.

Received: 30 June 2016 Accepted: 6 December 2016 References

1. Segal, IE: Observational validation of the chronometric cosmology: I. Preliminaries and the redshift-magnitude relation. Proc. Natl. Acad. Sci. USA72(7), 2473-2477 (1975)

2. Rosenblatt, M: A central limit theorem and a strong mixing condition. Proc. Natl. Acad. Sci. USA42, 43-47 (1956) 3. Berry, AE: The accuracy of the Gaussian approximation to the sum of independent variates. Trans. Am. Math. Soc.

49(1), 122-136 (1941)

4. Esseen, CG: On the Liapunaff limit of error in the theory of probability. Ark. Mat. Astron. Fys.28A(9), 1-19 (1942) 5. Parzen, E: On estimation of a probability density function and mode. Ann. Math. Stat.33, 1065-1076 (1962) 6. Woodroofe, M: Estimating a distribution function with truncated data. Ann. Stat.13, 163-177 (1985)

7. Stute, W: Almost sure representation of the product limit estimator of truncated data. Ann. Stat.21, 146-156 (1993) 8. Prakasa Rao, BLS: Berry-Esseen type bound for density estimation of stationary Markov processes. Bull. Math. Stat.17,

15-21 (1977)

9. Liang, HY, Una-Alvarez, J: A Berry-Esseen type bound in kernel density estimation for strong mixing censored samples. J. Multivar. Anal.100, 1219-1231 (2009)

10. Yang, W, Hu, S: The Berry-Esseen bounds for kernel density estimator under dependent sample. J. Inequal. Appl. 2012, 287 (2012)

11. Asghari, P, Fakoor, V, Sarmad, M: A Berry-Esseen type bound in kernel density estimation for a random left truncation model. Commun. Stat. Appl. Methods21(1), 115-124 (2014)

12. Asghari, P, Fakoor, V, Sarmad, M: A Berry-Esseen type bound for the kernel density estimator of length-biased data. J. Sci. Islam. Repub. Iran26(3), 256-272 (2015)

13. Lynden-Bell, D: A method of allowing for known observational selection in small samples applied to 3CR quasars. Mon. Not. R. Astron. Soc.155, 95-118 (1971)

14. He, S, Yang, GL: Estimation of the truncation probability in the random truncation model. Ann. Stat.26, 1011-1028 (1998)

15. Zhou, Y, Liang, H: Asymptotic normality forL1norm kernel estimator of conditional median underα-mixing

dependence. J. Multivar. Anal.73, 136-154 (2000)

16. Ould-Saïd, E, Tatachak, A: Strong consistency rate for the kernel mode estimator under strong mixing hypothesis and left truncation. Commun. Stat., Theory Methods38, 1154-1169 (2009)

17. Hall, P, Heyde, CC: Martingale Limit Theory and Its Application. Academic Press, New York (1980)

18. Yang, SC: Maximal moment inequality for partial sums of strong mixing sequences and application. Acta Math. Sin. Engl. Ser.23, 1013-1024 (2007)

19. Petrov, VV: Limit Theorems of Probability Theory. Oxford University Press, New York (1995)

20. Yang, SC, Li, YM: Uniformly asymptotic normality of the regression weighted estimator for strong mixing samples. Acta Math. Sin.49(5), 1163-1170 (2006)

21. Chang, MN, Rao, PV: Berry-Esseen bound for the Kaplan-Meier estimator. Commun. Stat., Theory Methods18(12), 4647-4664 (1989)

References

Related documents

THE INTERNATIONAL FESTIVAL OF WOMEN IN THE ARTS, Ny Carlsberg Glyptotek, DK FEMINISTISCHE KUNST INTERNATIONAAL, (Touring exhibition) Holland, Oosterport, Groningen;

In Model 4, we test Hypothesis 3, which predicts that the negative relationship between school ’ s percentage of single-parent families and performance can be explained by the

Phase transitions such as polymorph inter-conversion, formation of hydrates, desolvation of solvates, and conversion of the crystalline to amorphous form may occur during

It included all adults (above 18 years of age) who are household heads within the study area and store home grown maize. The sample size for households selected was determined

This study evaluated the use of WES in prenatal genetic diagnosis in fetuses with structural congenital anomalies (SCAs) detected on prenatal ultrasound.. Method: Thirty-three

of JARID1B promotes NSCLC cell proliferation, cell motil- ity, invasiveness, and tumorsphere formation in vitro, not only allude to a novel role of JARID1B in the modulation of EMT