Nothing Special   »   [go: up one dir, main page]

Academia.eduAcademia.edu

Moments of the Product-Limit Estimator Under Left-Truncation and Right-Censoring

2005, Communications in Statistics - Theory and Methods

Statistica Sinica 11(2001), 1069-1079 EDGEWORTH EXPANSIONS FOR THE PRODUCT-LIMIT ESTIMATOR UNDER LEFT-TRUNCATION AND RIGHT-CENSORING WITH THE BOOTSTRAP Yi-Ting Hwang Georgetown University School of Medicine Abstract: An Edgeworth expansion for the distribution function of the product-limit estimator of survival time under the left-truncation and right-censoring model is derived. This expansion gives more accurate approximations than the usual normal approximation from weak convergence. In addition, by constructing the bootstrap sample from left-truncation and right-censored data, the Edgeworth expansion for the bootstrap statistic is given, allowing a bootstrap base confidence interval with better coverage probability. Key words and phrases: Bootstrap, censoring, edgeworth expansion, truncation, U -statistic. 1. Introduction Censored or truncated data occur frequently in many fields such as epidemiology, astronomy and engineering life tests. For instance, consider a prevalent cohort study in epidemiology, which recruits a group of individuals at a specific time with a certain disease status and follows them over time. The variable of interest is the survival time defined as an individual’s age at death. Censoring occurs when an individual loses to follow-up, whereas truncation happens when an individual dies before the beginning of the follow-up study. See Hyde (1976), Tasi, Jewell and Wang (1987) and Wang (1991). Consider an infinite sequence of random vectors {(Xm , Tm , Cm ), m = 1, 2, . . .}, where the random variables X, T and C are nonnegative and independent with continuous distribution functions, FX , FT and FC , respectively. Suppose one can only observe the pair Z = min(X, C) and δ = I[X ≤ C], where I[A] denotes the indicator of the event A. Under this restriction, the random variable X is called right-censored by C. Furthermore, if (Z, T, δ) can be observed only when Z ≥ T , then the triple (Z, T, δ) is called a left-truncated and right-censored observation of X. For convenience, we denote the observable subsequence of (Zm , Tm , δm ), m = 1, 2, . . ., by (Ui , Vi , ηi ), i = 1, 2, . . . , where Ui = Zi , Vi = Ti , and ηi = δi if Zi ≥ Ti . The conditional distribution P[Z ≤ z, T ≤ t, δ = d|Z ≥ T ], 1070 YI-TING HWANG for z, t ∈ [0, ∞), d = 0 or 1 defines the left-truncation and right-censoring model (hereafter the TC model). If T ≡ 0 with probability 1, the TC model specializes to the right-censoring model; if C ≡ ∞ with probability 1, then the TC model becomes the random truncation model. The general problem is to make inference about the unknown distribution function FX (t) based on a sample of n independent identically distributed random vectors {(Uj , Vj , ηj ), j = 1, . . . , n} from {(Zm , Tm , δm ), m = 1, . . . , mn }, where n ≤ mn and mn is unknown. A product-limit estimator FX as defined in (1) is a well-known estimator for FX . The asymptotic properties of the PLE have been studied by Gijbels and Wang (1993) and He and Yang (2000). The weak convergence rate is O(n−1/2 ) which offers a confidence interval with coverage probability of O(n−1/2 ). However, under some regularity conditions, the bootstrap approximation is better than the normal approximation for a broad class of studentized statistics. Chen and Lo (1996) showed the bootstrap approximation for the studentized KaplanMeier estimator performs better than the normal approximation. Also, see Hall (1992) and Helmers (1991). The objective is to establish an Edgeworth expansion of a studentized PLE and find the bootstrap approximation of a studentized PLE under the TC model to improve the normal approximation. In Section 3 we establish the Edgeworth expansion of the studentized PLE, which provides an accuracy of o(n−1/2 ). However, since the PLE takes the product form, this is not done directly. By converting the target statistic into a U -statistic, we derive the Edgeworth expansion for the U -statistic first, then for the studentized PLE. In Section 4, we consider bootstrap samples obtained by simple random sampling with replacement from data. This allows us to obtain the expansions for the bootstrap of the studentized PLE. Using this expansion, we construct the coverage probability for the bootstrap approximation. Finally, simulations are conducted to provide numerical supports for the theoretical findings. 2. Assumptions and Notations Under left-truncation and right-censoring, the range of x for which FX (x) can be estimated needs to be carefully specified. Let aX and bX denote the lower and upper boundaries of the support of FX : aX = inf{z : FX (z) > 0} and bX = sup{z : FX (z) < 1}. Similar notation will be used for other distributions. It is well known that the non-parametric estimable range of FX is (aT , bZ ), where aT < bZ and bZ = min(bX , bC ). To ensure the finiteness of moments, we assume aT < min(aX , aC ) such that FT (max(aX , aC )) > 0. Moreover, we set aT = 0 to simplify the notation. EDGEWORTH EXPANSIONS FOR TC MODEL 1071 Let the survival function of FX be F̄X = 1 − FX . F̄C and F̄T are defined similarly. Denote marginal subdistributions by HU0 (u) = P[U ≤ u, η = 0] = α−1 HU1 (u) = P[U ≤ u, η = 1] = α−1 HV (v) = P[V ≤ v] = α−1  v 0  u 0  u 0 F̄X (z)FT (z)dFC (z), F̄C (z)FT (z)dFX (z), F̄C (s)F̄X (s)dFT (s), where α = P[T ≤ Z]. An important quantity for estimation is the coverage probability R(x) = P[V ≤ x ≤ U ] = α−1 FT (x)F̄C (x)F̄X (x). Let the corre 0 (s) = sponding empirical subdistributions of HU0 , HU1 and HV be given by H U n n −1 1 −1   I[U ≤ s, η = 0], H (s) = n I[U ≤ s, η = 1], H (s) = n i i i i V i=1 U i=1 n −1 I[Vi ≤ s]. Thus, the corresponding estimator for R(s) is Rn (s) = n i=1  V (s) − H  0 (s−) − H  1 (s−) where, for any function n−1 ni=1 I[Vi ≤ s ≤ Ui ] = H U U h(x), h(x−) denotes the left-continuous version of h(x). Then the well-known product-limit estimator (PLE) of FX is given by FX (t) = 1 −  1− s≤t  1 (s)  ∆H U , Rn (s) t ∈ (0, bZ ), (1)  1 (s) is the difference H  1 (s) − H  1 (s−). Here we use the convention where ∆H U U U that 0/0 = 0. Let the cumulative hazard function of FX be denoted as  t ΛX (t) = 0 dFX (s) = − ln F̄X (t). 1 − FX (s−) (2) Replacing FX by the PLE FX in (2), we obtain an estimator − ln F̄ X (t) for the cumulative hazard function. Since F̄ X is complicated, we use the cumulative hazard estimator as an auxiliary estimator for deriving an Edgeworth expansion for the PLE. As discussed in Hwang (2000), FX (t) = 0 with positive probability. To avoid this difficulty, we partition the sample space Ω as follows: Ω0 = {ω ∈  V − HV |, |H  0 − H 0 |, |H  1 − H 1 |) < γ} and Ω1 = Ω − Ω0 , Ω : sup0≤s≤t max(|H U U U U where for a fixed t, τ ∈ (0, t] is chosen such that Θ = P[V ≤ τ, U ≥ t] > 0 and γ is a real number with 0 < γ < Θ /3 and 6γ(Θ − 3γ)−2 < 1. Since the coverage probability R(x) is not necessarily monotone, it is necessary to introduce Θ to compute the probability bound. Clearly, Θ ≤ R(s), for any τ ≤ s ≤ t. Then,  (t) + ln F̄ (t)| < ∞. Also, from the Dvoretzky, for ω ∈ Ω0 , we have | − ln F̄ X X Kiefer and Wolfowitz (1956) inequality (DKW inequality hereafter), we have P[Ω1 ] = o(n−k ). Thus, we focus discussion on the subspace Ω0 . 1072 YI-TING HWANG 3. The Edgeworth Expansion for the Studentized PLE √  (t) + ln F̄ (t)) be written as σ 2 = Let the asymptotic variance of n(− ln F̄ X X 0  t −2 1 0 R dHU (see Gijbel and Wang (1993)) and the empirical variance estimator of 02 = σ02 be σ t 0  1 . Let Φ(x) and φ(x) be the standard normal distribution Rn−2 dH U function and the standard normal density function, respectively. The following theorem derives the Edgeworth expansion for the cumulative hazard estimator: Theorem 3.1. We have   sup P x  n1/2  − ln F̄ X (t) + ln F̄X (t) ≤ x − Ψn (x) = o(n−1/2 ) 0 σ (3) uniformly in x, where Ψn (x) = Φ(x) + n−1/2 φ(x)[κ1 x2 + κ2 + σ02 (2n1/2 )−1 ], and 1 κ1 = 3 3σ0 κ2 = 1 6σ03 +  t 0  t 0 1 2σ03 α R−3 (u)dHU1 (u), (4) R−3 (u)dHU1 (u)  t 0 R−2 (u)F̄C (u)  u 0 R−2 (s)FT (s)dHU1 (s)dFX (u). (5) Now we are ready to present the main theorem in this section. Let the √  2 2 2 asymptotic variance of n(F̄ X (t) − F̄X (t)) be σ = F̄X (t)σ0 and the empirical 2  σ 2  2 = F̄ variance estimator of σ 2 be σ X 0 . Theorem 3.2. We have   sup P x  n1/2   F̄ X (t) − F̄X (t) ≤ x − Ψn (x) = o(n−1/2 )  σ uniformly in x, where Ψn (x) = Φ(x) + n−1/2 φ(x) κ1 x2 + κ2 , σ0−3 3  t σ −3 κ2 = − 0 6  t κ1 = − − 0 0  σ0−3 t 2α 0 R−3 dHU1 + σ0 , 2 R−3 dHU1 − σ0 2 R −2  u (u)F̄C (u) 0 (6) (7) R−2 (s)FT (s)dHU1 (s)dFX (u). (8) EDGEWORTH EXPANSIONS FOR TC MODEL 1073 4. The Bootstrap Statistic From the normal approximation, a confidence interval of FX (t) can be constructed with a coverage probability accurate to O(n−1/2 ). By means of the bootstrap, we show that the coverage probability is accurate to order o(n−1/2 ). A bootstrap sample is obtained by simple random sampling with replacement from {(Ui , Vi , ηi ), i = 1, . . . , n}. Let {(Ui∗ , Vi∗ , ηi∗ ), i = 1, . . . , n} denote the bootstrap sample. The symbol ∗ represents statistics associated with the bootstrap sample. For instance, P∗ is the probability measure on the bootstrap sample,  1∗ (u) = H U n 1 I[Ui∗ ≤ u, ηi∗ = 1], n i=1 g∗ (Ui∗ , Vi∗ , ηi∗ ) = Rn−1 (Ui∗ )ηi∗ I[0 ≤ Ui∗ ≤ t]  t +  Now the process 0  1 (s). Rn−2 (s)(I[s ≤ Vi∗ ] − I[s < Ui∗ ])dH U   1∗ (s), s ∈ [0, ∞) H U is the empirical process with the parent  1 (s), s ∈ [0, ∞)}. distribution {H U Let Ψ∗n (x) = Φ(x) + n−1/2 φ(x)(κ∗1 x2 + κ∗2 ), where κ∗1 and κ∗2 are the corresponding bootstrap estimates of κ1 and κ2 as defined in (7) and (8). Note that κ∗1 and κ∗2 depend only on the sample (Ui , Vi , ηi ), i = 1, . . . , n, and not on the bootstrap samples. The following theorem gives the bootstrap accuracy for the studentized PLE. The proof is similar to that of Theorem 3 in Helmers (1991) and is there fore omitted . Theorem 4.1. For ω ∈ Ω0 we have    n1/2  ∗  (t) ≤ x − Ψ∗ (x) = o(n−1/2 ), (t) − F̄ F̄ X X n ∗ σ    n1/2  ∗  (t) ≤ x − Q  n (x) = o(n−1/2 ), (t) − F̄ F̄ X X ∗ σ sup P∗ x sup P∗ x  (t) − F̄ (t)) ≤ x].  n (x) = P[√nσ  −1 (F̄ where Q X X From the Edgeworth expansion for the bootstrap statistic, we can now construct confidence intervals with better coverage probabilities. Let zα = Φ−1 (α). The normal approximation yields the following one-sided confidence interval for  n−1/2 ). It is easy to see that by (6), we have F̄X (t): (−∞, F̄ X − zα σ  n−1/2 = 1 − α + n−1/2 φ(zα )(κ1 zα2 + κ2 ) + o(n−1/2 ). P F̄X (t) ≤ F̄ X (t) − zα σ Therefore, the error in the coverage probability for the normal based confidence  n . According to interval is of order O(n−1/2 ). Let qα denote the α-quantile of Q 1074 YI-TING HWANG the inversion formula for the Edgeworth expansion (Hall (1992, p.88)), we have zα = qα + n−1/2 (−κ1 zα2 − κ2 ). Thus, the error in approximating the quantile zα by qα is of order O(n−1/2 ). √  ∗ (t) − F̄  (t)).  ∗ )−1 (F̄ Let qα∗ be the α-quantile of the distribution of n(σ X X Then, the following theorem shows that the coverage probability for the bootstrap -based confidence interval is accurate to o(n−1/2 ). Also, the error in estimating the quantile qα∗ by qα is of order o(n−1/2 ). Theorem 4.2. For fixed 0 < α < 1, we have qα∗ = qα + o(n−1/2 ) a.s. and  (t) − q ∗ σ −1/2 P F̄X (t) ≤ F̄ = 1 − α + o(n−1/2 ). X α n (9) The proof, based on a standard delta method, can be taken from Theorem 7 in Chen and Lo (1996). Example 4.1. To examine the result of Theorem 4.2, consider the distribution functions FX (s) = 1 − exp(−(s − a)), FC (s) = 1 − exp(−βC (s − a)) and FT (s) = 1 − exp(βT s), where a constant a is chosen so that FT (a) > 0. Here we set a = 0.01. The truncation probability is 0.351. The coverage probabilities of onesided confidence intervals constructed on normal approximations and bootstrap approximations are shown in Table 1 for three fixed time points, t = 0.7, t = 0.6 and t = 0.5. All three bootstrap approximations perform better than the normal approximatioin at all nominal levels. At time t = 0.7 and nominal level 0.975, the improvement for the bootstrap approximation is 100%. However at time 0.5, the improvement does not look as dramatic. The normal and bootstrap approximations are rougher at t = 0.5 due to fewer observations in calculating the PLE. Table 1. Coverage probability of confidence interval∗ Time t = 0.7 Time t = 0.6 Time t = 0.5 n∗∗ = 28 n = 25 n = 29 Nominal Normal Bootstrap Normal Bootstrap Normal Bootstrap 0.975 0.932 0.975 0.920 0.974 0.912 0.973 0.95 0.888 0.949 0.872 0.935 0.863 0.921 0.90 0.816 0.885 0.797 0.850 0.788 0.828 0.85 0.753 0.799 0.735 0.768 0.728 0.742 *Sample size=50. The bootstrap approximations are based on 1000 repetitions. **n represents the truncated sample size. Remark 4.1. During the revision, the author was made aware of a recent unpublished manuscript by Drs. Wang and Jing (2000) which addresses the same EDGEWORTH EXPANSIONS FOR TC MODEL 1075 problems as presented in this article. Both papers use similar methodologies; however, the results on Edgeworth expansions for the studentized PLE are different. It is the author’s observation that while implicating Theorem 1.2 in Bickel, Götze and van Zwet (1986), the coefficient κ3 in Lemma 5.5 in Wang and Jing is derived incorrectly. 5. Extensions The independence assumption on (X, T, C) can be further relaxed by assuming that X is independent of the pair (T, C), with possible dependence between the random variables T and C. Acknowledgements The author would like to thank her dissertation advisor Professor Grace L. Yang. The author is also grateful for valuable comments from the Associate Editor and the referee. This research was partially supported by Agency for HealthCare Research and Quality Grant #HS08395 and Contract DAMD 17-96-C-6069 from the Department of the Army. Appendix A. Proofs  (t) + ln F̄ (t), we need To derive the U -statistic representation for − ln F̄ X X the following notation. For 1 ≤ j, k ≤ n, let B1 (Uj , Uk , Vk ) = R−2 (Uj )I[0 ≤ Uj ≤ t](I[Uj ≤ Vk ] − I[Uj < Uk ]),  t B2 (Uj , Vj , Uk , Vk ) = 0  R−3 (s) i=j,k (I[s ≤ Vi ] − I[s < Ui ])dHU1 (s). In particular, when j = k, it is easy to see that 1 B1 (Uj ) = B1 (Uj , Vj , Uj , Vj ) = − R−2 (Uj )I[0 ≤ Uj ≤ t], 2 B2 (Uj , Vj ) = B2 (Uj , Vj , Uj , Vj ) = −  t 0 R−3 (s)(I[s ≤ Vj ] − I[s < Uj ])dHU1 (s). To simplify the notation, set Ui = (Ui , Vi , ηi ) and let g(Uj ) = R −1 (Uj )I[0 ≤ Uj ≤ t] +  t 0 R−2 (s)(I[s ≤ Vj ] − I[s < Uj ])dHU1 (s), ψ(Uj , Uk ) = B1 (Uj , Uk , Vk )ηj + B2 (Uj , Vj , Uk , Vk ) −E[B1 (Uj , Uk , Vk )ηj + B2 (Uj , Vj , Uk , Vk )|Uj , Vj , ηj ], h(Uj , Uk ) = g(Uj ) + g(Uk ) + ψ(Uj , Uk ) + ψ(Uk , Uj ). 1076 YI-TING HWANG For ω ∈ Ω0 , Lemma 4.1 in Hwang (2000) yields  (t) + ln F̄ (t) = U + Rem + − ln F̄ X X n V − H 0 − H  1 )(s), Un = n−2 where Rn (s) = (H U U  t Rem= 0 +  1 −H 1 ]+ R−3 [Rn −R]2 d[H U U ∞   k  k=3 i=2 (−1)k−1 k−1 +n−2 n   n−i+1 ∞   (−1)k k=3   t k i  0 i<j  t 0 (10) h(Ui , Uj ) and 1 R−(k+1) [Rn −R]k dH U 1 R−k (Rn − R)k−i dH U 2g(Ui ) + B1 (Ui )ηi + B2 (Ui , Vi ) − i=1 σ02 , 2n  (11)  σ02  . 2 (12) (13) We can show that √ P[ nσ0−1 |Rem| > (log n)−1 n−1/2 ] = o(n−1/2 ) (14) using Lemma 3.2 in Hwang (2000) and the DKW inequality. Details are in Lemma 5.1.2 in Hwang (1999). Since Un is a U -statistic of order n−2 with a symmetric kernel h, Un can   be expressed as Un = n−2 {(n − 1) ni=1 g(Ui ) + i<j [ψ(Ui , Uj ) + ψ(Uj , Ui )]}, which has the form of (1.5) in Bickel, Götze and van Zwet (1986). From their Theorem 1.2, we obtain the following lemma. The proof is similar to that of Lemma 1 in Chang (1991). Note that, because of the dependence between U and V , the coefficient K3 is different from that in Chang’s Lemma 1.  √    Lemma A.1 We have supx P[ nσ0−1 Un ≤ x] − Gn (x) = o(n−1/2 ), Gn (x) = Φ(x) − K3 (6n1/2 )−1 φ(x)(x2 − 1) and K3 = σ0−3 +  t 0 3 σ03 α R−3 (u)dHU1 (u)  t 0 R−2 (s)F̄C (s)  s 0 R−2 (u)FT (u)dHU1 (u)dFX (s). (15)  −2 dH 1 − 02 − σ02 = 0t Rn Proof of Theorem 3.1. To use Lemma A.1, write σ U  t −2  n 1 −1 2 R dH = n f (U )+ξ , where f (U ) = −2 B (U )η − B (U , U )+σ i 1 i 1 i i 2 i j i=1 0 U  0  t −3 t  1 − H 1 ) + R−2 R−3 (Rn − R)2 (R + 2Rn )dH 1 . and ξ1 = −2 0 R (Rn − R)d(H U U U 0 n Thus, we have n 1  σ0 =1− f (Ui ) + ξ2 , 0 σ 2nσ02 i=1 (16) EDGEWORTH EXPANSIONS FOR TC MODEL 1077 0 − σ0 )2 (σ 0 + 2σ0 )(2σ 0 σ02 )−1 . By (10) and (16), we where ξ2 = −ξ1 (2σ02 )−1 + (σ have σ0 n1/2 − ln F̄ X (t) + ln F̄X (t) = ζ + ξ3 − 1/2 , (17) 0 σ 2n where ζ= ξ3 = n1/2 2  h(Ui , Uj ) 2σ0 n2 i<j 1− n 1  f (Ui ) , 2nσ02 i=1 n  n1/2 ξ2 2  n1/2 1 σ0 ξ 2 h(U , U ) + Rem + f (U ) − . i j i 2 0 2σ0 n i<j σ 2n3/2 σ0 i=1 2n1/2 The quantity ζ can be rewritten as  −1 n − 1 n1/2 n ζ= n 2σ0 2  h(Ui , Uj ) − i<j  1 E[g(U )f (U )] + Rem1 , (18) i i nσ02 where h(Ui , Uj ) = h(Ui , Uj ) − (2σ02 )−1 [g(Ui )f (Uj ) + g(Uj )f (Ui )] and Rem1 = −  1 2n5/2 σ03 i<j n  n−1 − 5/2 3 2n σ0 − 1 2n5/2 σ03 {ϕ(Ui , Uj )(f (Ui )+f (Uj ))−g(Ui )f (Uj )+g(Uj )f (Ui )} {g(Ui )f (Ui ) − E[g(Ui )f (Ui )]} i=1 n   f (Ui ) i=1   ϕ(Uk , Um ) . k<m k=i=m  Clearly, i<j h(Ui , Uj ) is a U -statistic of degree 2 with an expected value of zero. Also, we have E[h(Uj , Uk )|Uj ] = g(Uj ) and E[g(Uj )] = E[f (Uj )] = 0 from Lemma 3.1 in Chang and Hwang (2000). Thus, by Lemma A.1 and Lemma 2 in Chang (1991), we have P E[g(Ui )f (Ui )] n1/2  h(Ui , Uj ) − ≤x σ0 n2 i<j nσ02 = Φ(x) + n−1/2 φ(x)(κ1 x2 + κ2 )+o(n−1/2 ), where κ1 and κ2 are as defined in (4) and (5). To finish the proof, we must show that the error term Rem1 and ξ3 are of order o(n−1/2 ). Calculations for the moments of Rem1 can be found in Callaert and Veraverbeke (1981). Then, applying Chebyshev’s inequality, one can show that (19) P[|Rem1 | > (n log n)−1/2 ] = o(n−1/2 ). 1078 YI-TING HWANG Moreover, we have P[|ξ3 | > (n log n)−1/2 ] = o(n−1/2 ). (20) The proof applies Hoeffding’s inequality, Bernstein’s inequality (Serfling (1980, p.85)) and (14). Details are described in Lemma A.5.2. in Hwang (1999). The proof is therefore complete. Proof of Theorem 3.2. From a Taylor expansion and (10), we obtain n1/2 n1/2   (t) + ln F̄ (t) − ln F̄ F̄ X (t) − F̄X (t) = − X X  0 σ σ − 1+ n 1  g(Ui ) 2n i=1 n1/2  (t) + ln F̄ (t) × Rem , − ln F̄ X X 2 0 σ (21) where | − 1| ≤ exp(− ln F̄ X (t) + ln F̄X (t)) and   n  1   ϕ(Ui , Uj )− g(Ui ) +Rem+ −ln F̄ X (t)+ln F̄X (t) Rem2 = 2 2n i<j 6 i=1 2 + σ02 . 2n Applying (14), Lemma 3.2 in Hwang (2000) and Lemma 3.1 in Chang and Hwang (2000), the quantity Rem2 can be shown to have a second moment of order O(n−2 ). Therefore, combining Theorem 3.1 and the fact that P[|Rem2 | > (n log n)−1/2 ] = o(n−1/2 ), the second term of the left side of (21) is clearly negligi ble. From (17), the first term in (21) can be rewritten as−ζ(1+(2n)−1 ni=1 g(Ui )) √ −1  + σ0 (2 n) + Rem3 , where Rem3 = −ξ3 (1+(2n)−1 ni=1 g(Ui )) + σ0 (4n3/2 )−1 n −1/2 ] i=1 g(Ui ). By (20) and Bernstein’s inequality, we have P[|Rem3 | > (n log n) −1/2 ). To complete the proof, it suffices to show that −ζ(1 + (2n)−1 = o(n √ −1 n has the same Edgeworth expansion as Ψn (x), as dei=1 g(Ui )) + σ0 (2 n) fined in (6). From the proof of (18) and (19), ζ can be reexpressed as (n3/2 σ0 )−1  2 −1 −1/2 ), where the first term can i<j {h(Ui , Uj ) − (nσ0 ) E[g(Ui )f (Ui )]} + o(n be represented as a U -statistic by using the technique employed in (18). We complete the proof by applying Lemma A.1 and Lemma 2 in Chang (1991). References Bickel, P. J., Götze, F. and van Zwet, W. R. (1986). The Edgeworth expansion for U-statistics of degree two. Ann. Statist. 14, 1463-1484. Callaret, H. and Veraverbeke, N. (1981). The order of the normal approximation for a studentized U -statistics. Ann. Statist. 9, 194-200. Chang, D. C. and Hwang, Y. T. (2000). A Berry-Esséen bound for the product-limit estimator under left-truncation and right-censoring. To appear in Appl. Anal. Chang, M. N. (1991). Edgeworth expansion for the Kaplan-Meier estimator. Commun. Statist. Theory Meth. 20, 2479-2494. EDGEWORTH EXPANSIONS FOR TC MODEL 1079 Chen, K. and Lo, S. H. (1996). On bootstrap accuracy with censored data. Ann. Statist. 24, 569-595. Dvoretzky, A., Kiefer, J. and Wolfowitz, J. (1956). Asymptotic minimax character of the sample distribution function and of the classical multinomial estimator. Ann. Math. Statist. 27, 642-669. Gijbels, I. and Wang, J. L. (1993). Strong representations of the survival function estimator for truncated and censored data with applications. J. Mult. Anal. 47, 210-229. Hall, P. (1992). The Bootstrap and Edgeworth Expansion. Springer-Verlag, New York. Helmers, R. (1991). On the Edgeworth expansion and the bootstrap approximation for a studentized U -statistics. Ann. Statist. 19, 470-484. He, S. and Yang, G. L. (2000). On the strong convergence of the product-limit estimator and its integrals under censoring and random truncation. Statist. Probab. Lett. 49, 235-244. Hwang, Y. T. (1999). Finite moments for the left-truncation and right-censoring model. Ph.D. Dissertation, Dept. of Mathematics, University of Maryland. Hwang, Y. T. (2000). Moments of the product-limit estimator under left-truncation and rightcensoring. Submitted for publication. Hyde, J. (1977). Testing survival under right censoring and left truncaton. Biometrika 64, 225-230. Serfling, R. J. (1980). Approximation Theorems of Mathematical Statistics. John Wiley, New York. Tsai, W. Y., Jewell, N. P. and Wang, W. C. (1987). A note on the product-limit estimtor under right censoring and left truncation. Biometrika 74, 883-886. Wang, M. C. (1991). Nonparametric estimation from cross-sectional survival data. J. Amer. Statist. Assoc. 86, 130-143. Wang, Q. H. and Jing, B. Y. (2000). Edgeworth expansion and bootstrap approximation for studentized product-limit estimator with truncated and censored data. Unpublished. Department of Oncology, Georgetown University School of Medicine, Lombardi Cancer Center, Washington, DC, U.S.A. E-mail: yh4@gunet.georgetown.edu (Received April 2000; accepted April 2001)