Professional Documents
Culture Documents
condition
Zohra Guessoum1 and Elias Ould-Sad2
1
Faculte de Mathematiques, Univ. des Sci. et Tech. Houari Boumedienne, BP 32, El
Alia, 16111, Algeria.
e-mail: z0guessoum@hotmail.com
2
L.M.P.A. J. Liouville, Univ. du Littoral Cote dOpale, BP 699, 62228 Calais, France.
e-mail: ouldsaid@lmpa.univ-littoral.fr(corresponding author)
Contents
1 Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
2 Definition of estimators . . . . . . . . . . . . . . . . . . . . . . . . 3
3 Assumptions and main result . . . . . . . . . . . . . . . . . . . . . 5
4 Simulations Study . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
5 Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
1. Introduction.
Cox [43] established the same result with a slight difference on the rate.
Andrews [1] provides a comprehensive set of results concerning the uniform
almost sure convergence. Masry [36] derived sharp rates for the same kind
of convergence, but confined attention to the case of bounded regression.
Our goal is to establish the strong uniform convergence with rate for the
kernel regression estimate under -mixing condition in random censorship
models. For this kind of model, Cai [5, 7] established the asymptotic proper-
ties of the Kaplan-Meier estimator. The strong convergence of a hazard rate
estimator was examined by Lecoutre and Ould-Sad [31] while Liebscher [35]
derive a rate uniform for the strong convergence of kernel density and hazard
rate estimators. His result represents an improvement of that given in Cai
[6]. The consistency results concerning the nonparametric estimates of the
conditional survival function introduced by Beran [2], Dabrowska [16, 17]
in the iid case, were extended by Lecoutre and Ould-Sad [32] to the strong
mixing case. We point out that, for the independent case, the behavior of
the regression function under censorship model has been extensively stud-
ied. We can quote Carbonez et al. [12], Kohler et al. [28] and Guessoum
and Ould-Sad [23]. However, few papers deal with the regression function
under censoring in the dependent case.
To this end, we were interested in extending the result of Guessoum and
Ould-Sad [23] from the iid to the dependent case. The paper is organized as
follows: In Section 2 we give some definitions and notations under the cen-
sorship model of the regression function and strong-mixing process. Section
3 is devoted to the assumptions and main result. In Section 4, some simu-
lations are drawn to lend further support to our theoretical results. Proof
with auxiliary results are relegated to Section 5.
2. Definition of estimators
with f, (x, y) being the joint density of (X, Y ) and () the density function
of the covariates.
Now, it is well known that the kernel estimator of the regression function
r1,n (x)
mn (x) =:
n (x)
where
n n
1 X i Ti x Xi 1 X x Xi
r1,n (x) = Kd and n (x) = Kd .
nhdn i=1 G( Ti ) hn nhdn i=1 hn
(3)
In practice, G is usually unknown, we replace it by the corresponding Kaplan-
Meier [27] estimator (KME) Gn defined by
1i 1I{Yi t}
Qn
i=1 1 , if t < Y(n) ,
Gn (t) = ni+1
0, if t Y(n) .
The properties of the KME for dependent variables can be found in Cai
r (x)
[5, 7]. Then a feasible estimator of m(x) is given by: mn (x) = 1,n
n (x)
where
n
1 X i Ti x Xi
r1,n (x) = Kd (4)
nhdn i=1 Gn ( Ti ) hn
under study. We point out that since Y can be a lifetime we can suppose it
bounded. We put ktk = dj=1 |tj | for t IRd .
P
A3) The mixing coefficient is such that (n) = O(n ) for some >
p + p2 + 3(d 1) where p = (4+d)+d
p
2 .
A4) The function r1 () definedZ in (1) is continuously differentiable.
A5) The function r 2 (x) := y 2 fX,Y (x, y)dy is continuously differen-
d
IR
tiable.
A6) D > 0 such that supu,vC |ij (u, v) (u)(v)| < D where ij is the
joint distribution of (Xi , Xj ).
A7) > 0, c1 > 0, c2 > 0, such that
(3) d
+d
c1 n d[(+1)+2+1] hdn c2 n 1 .
A8) The marginal density (.) is continuously differentiable and there exists
> 0 such that (x) > x C.
Remark 3.2 The rate obtained here is slightly different from that obtained
nq o
log logn 2
by Guessoum and Ould-Sad [23] in the independent case, which is O max nhn , hn
for d = 1. Their result can easily be generalized for higher dimensional co-
IRd , by adapting
variate, ie X nq o their Assumptions A1 and A2 to obtain
log logn 2
the rate O max nhd
, hn .
n
This is the optimal rate obtained by Liebscher [34] in the uncensored case.
4. Simulations Study
1.5 2 2
continuous curve=True curve continuous curve=True curve continuous curve=True curve
dashed curve=Estimated curve dashed curve=Estimated curve dashed curve=Estimated curve
1.5 1.5
1
1 1
0.5
0.5 0.5
0
0 0
0.5
0.5 0.5
1
1 1
1.5
1.5 1.5
2 2 2
2 1.5 1 0.5 0 0.5 1 2 1.5 1 0.5 0 0.5 1 1.5 2 2 1.5 1 0.5 0 0.5 1 1.5 2
We notice that the quality of fit increases with n (see Figure 1).
Yi = sin( Xi ), sinus case, (5)
2
5 2
Yi = X 2, parabolic case. (6)
12 i+1
5 2 2
Then we have m(x)= sin( 2 x) for (5) and m(x)= 12 x + 5
12 (1 2 ) 2
for (6).
1 1 1
0 0 0
1 1 1
2 1.5 1 0.5 0 0.5 1 1.5 2 2 1.5 1 0.5 0 0.5 1 1.5 2 2 1.5 1 0.5 0 0.5 1 1.5 2
2 2 2
5 2 2
Fig 3. m(x) = 12
x + 5
12
(1 2 ) 2, with n = 50, 100 and 300, respectively.
Figures 2 and 3 show that the quality of fit for the non linear model is as
good as in the linear model.
5. Proofs
= IE [Y1 |X1 = u]
= m(u).
since r1 = m.
A Taylor expansion gives
r1 r1
r1 (x hn t) r1 (x) = hn (t1 (x ) + ... + td (x ))
x1 xd
where x is between x hn t and x. Then
Z
sup |IE(r1,n (x)) r1 (x)| = sup Kd (t) [r1 (x hn t) r1 (x)] dt
xC xC IRd
r1 r1
Z
hn sup Kd (t)(t1 (x ) + ... + td (x ))dt .
xC d
IR x 1 x d
where Sn2 =
PP
i j |cov(Ui , Uj )| .
Lemma 5.3 Under Assumptions A1-A7, we have
s !
log n
sup |r1,n (x) IEr1,n (x))| =O a.s as n .
xC nhdn
It is obvious that
n
X
i (x) = r1,n (x) IEr1,n (x)).
i=1
i (x), we have clearly |i (x)|
Writing i (x)i (xk ) = i (x)+|i (x )|.
k
Then,
n n
( )
1X i |Ti | 1 x Xi xk Xi
X
sup
i (x) sup Kd Kd
xC i=1
xC n i=1 G(Ti ) hdn hn hn
1 |T1 | 1 x X1 xk X1
+ sup IE d
Kd Kd .
xC G(T1 ) hn hn hn
Sn2 =
X X
|cov(Ui , Uj )| + nV ar(U1 ).
i j, i6=j
We have
" #
12 T12
2 xk X1 1 T1 xk X1
2
V ar(U1 ) = IE Kd IE Kd =: I1 I2 .
G2 (T1 ) hn G(T1 ) hn
Using the conditional expectation properties and a change of variables, we
get
" #
12 T12
2 xk X1
I1 = IE Kd
G2 (T1 ) hn
" !#
xk X1 12 T12
= IE Kd2 IE |X1
hn G2 (T1 )
hdn
Z
Kd2 (t)r2 (xk hn t)dt.
G(F ) IR d
I1 = O(hdn ).
i Ti xk Xi i Ti xk Xi
|cov(Ui , Uj )| = |IEUi Uj )| = IE Kd IE Kd
G(Ti ) hn G(Ti ) hn
" !#)
j Tj xk Xj j Tj xk Xj
Kd IE Kd
G(Tj ) hn G(Tj ) hn
!
Yi xk Xi Yj xk Xj
IE Kd Kd
G(Yi ) hn G(Yj ) hn
!
Yi xk Xi Yj xk Xj
IE Kd IE Kd
G(Yi ) hn G(Yj ) hn
Z Z
Ch2d
n Kd (z)Kd (t) [ij (xk zhn , xk thn ) i (xk zhn )j (xk thn )] dz dt.
IRd IRd
Assumption A6 gives
|cov(Ui , Uj )| = O(h2d
n ). (7)
h2d
XX XX
C n + C (|i j|)
|ij|>wn 0<|ij|wn
C nh2d 2
n wn + Cn (wn ) .
h i
1 1
Now choosing wn = hdn
+1, we have Sn2 O(nhdn )+Cn2 hdn
. Assump-
tion A3 and the right part of Assumption A7 yield n2 ( h1d ) = O(nhdn ) .
n
So
Sn2 = O(nhdn ).
Finally, we have
Sn2 = Sn2 + nV ar(U1 ) = O(nhdn ).
Then, for > 0, applying Lemma 5.2, we have
( n ) ( n )
X X
k d
IP i (x ) > = IP Ui > nhn
i=1 i=1
! q +1
2 nhdn 2
q
1
C 1+C + nCq (9)
.
q nhdn
s
log n
If we replace by 0 for all 0 > 0 in (9), we get
nhdn
( n s ) ! q +1
X
k log n 20 log n 2
q
IP i (x ) > 0 C 1+C +nCq 1 .
nhdn
q
i=1
q 0 nhdn log n
(10)
By choosing q = (log n)1+b (b > 0), (10) becomes
( n s ) +1
X log n
q +1
k C20 1
IP i (x ) > 0 Cn + nCq nhdn log n 2
i=1
nhdn 0
d(+1)
2 (+1) +1
CnC0 + C0 (log n)(1+b) n1 2 hn 2
.
1
d(1+ 2 ) d 2
CM hn n 2 C0
d(+1) d
(+1) +1 d 2 d
+ M C0 (log n)(1+b) n1 2
+ 2
hn 2
(+1)
=: CM J1 + M C0 J2 . (11)
(3) (+1)+2+1
1 +1 d
+ 2 d( )
J2 C (log n)(1+b) n 2 n 2 2
(+1)+2+1 1
1d( )
C (log n)(1+b) n 2 .
sup |(x) n (x)| sup |n (x) IE (n (x))| + sup |IE (n (x)) (x)| .
xC xC xC
n
1 1 X x Xi
sup Gn (t) G(t) |Yi | Kd .
Gn (F )G(F ) tF nhdn i=1
hn
In the same way as for Theorem 2 of Cai (2001), it can be shown under A3
that s
log log n
sup Gn (t) G(t) = O a.s. Furthermore, from the defini-
tF n
tion of n (x), Lemma 5.4, Assumptions A1, A2 and A8, and the fact that
Y is bounded we get the result.
Proof of Theorem 3.1: We have
r1,n (x) r1,n (x) r1,n (x) IEr1,n (x))
sup |mn (x) m(x)| sup +
xC xC n (x) n (x) n (x) n (x)
IE(r1,n (x)) r1 (x) r1 (x) r1 (x)
+ +
n (x) n (x) n (x) (x)
(
1
sup |r1,n (x) r1,n (x)| + sup |r1,n (x) IEr1,n (x))|
inf n (x) xC xC
)
1
+ sup |IE(r1,n (x) r1 (x))| + sup |r1 (x)| sup |(x) n (x)| . (13)
xC xC xC
The kernel estimator n (x) is almost surely bounded away from 0 because
of Lemma 5.4 and the fact that the second part of Assumption A8.
Then, (13) in conjunction with Lemmas 5.1, 5.3, 5.4 and 5.5 we conclude
the proof.
References