The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.

Marginal modelling of clustered survival data

Klaus Holst & Thomas Scheike

2024-02-16

Overview

A basic component for our modelling of multivariate survival data is that many models are build around marginals that on Cox form. The marginal Cox model can be fitted efficiently in the mets package, in particular the handling of strata and robust standard errors is optimized.

The basic models assumes that each subject has a marginal on Cox-form \[ \lambda_{g(k,i)}(t) \exp( X_{ki}^T \beta). \] where \(g(k,i)\) gives the strata for the subject.

We here discuss and show how to get robust standard errors of

and how to do goodness of fit test using

First we generate some data from the Clayton-Oakes model, with \(5\) members in each cluster and a variance parameter at \(2\)

 library(mets)
 options(warn=-1)
 set.seed(1000) # to control output in simulatins for p-values below.
 n <- 1000
 k <- 5
 theta <- 2
 data <- simClaytonOakes(n,k,theta,0.3,3)

The data is on has one subject per row.

Now we fit the model and produce robust standard errors for both regression parameters and baseline.

First, recall that the baseline for strata \(g\) is asymptotically equivalent to \[\begin{align} \hat A_g(t) - A_g(t) & = \sum_{k \in g} \left( \sum_{i: ki \in g} \int_0^t \frac{1}{S_{0,g}} dM_{ki}^g - P^g(t) \beta_k \right) \end{align}\] with \(P^g(t) = \int_0^t E_g(s) d \hat \Lambda_g(s)\) the derivative of \(\int_0^t 1/S_{0,g}(s) dN_{\cdot g}\) wrt to \(\beta\), and \[\begin{align} \hat \beta - \beta & = I(\tau)^{-1} \sum_k ( \sum_i \int_0^\tau (Z_{ki} - E_{g}) dM_{ki}^g ) = \sum_k \beta_{k} \end{align}\] with \[\begin{align} M_{ki}^g(t) & = N_{ki}(t) - \int_0^t Y_{ki}(s) \exp( Z_{ki} \beta) d \Lambda_{g(k,i)}(t), \\ \beta_{k} & = I(\tau)^{-1} \sum_i \int_0^\tau (Z_{ki} - E_{g}) dM_{ki}^g \end{align}\] the basic 0-mean processes, that are martingales in the iid setting, and \(I(t)\) is the derivative of the total score, \(\hat U(t,\beta))\), with respect to \(\beta\) evaluated at time \(t\).

The variance of the baseline of strata g is estimated by \[\begin{align} \sum_{k \in g} ( \sum_{i: ki \in g} \int_0^t \frac{1}{S_{0,g(k,i)}} d\hat M_{ki}^g - P^g(t) \beta_k )^2 \end{align}\] that can be computed using the particular structure of \[\begin{align} d \hat M_{ik}^g(t) & = dN_{ik}(t) - \frac{1}{S_{0,g(i,k)}} \exp(Z_{ik} \beta) dN_{g.}(t) \end{align}\]

This robust variance of the baseline and the iid decomposition for \(\beta\) is computed in mets as:

   out <- phreg(Surv(time,status)~x+cluster(cluster),data=data)
   summary(out)
#> 
#>     n events
#>  5000   4854
#> coeffients:
#>   Estimate     S.E.  dU^-1/2 P-value
#> x 0.287859 0.028177 0.028897       0
#> 
#> exp(coeffients):
#>   Estimate   2.5%  97.5%
#> x   1.3336 1.2619 1.4093
   # robust standard errors attached to output
   rob <- robust.phreg(out)

We can get the iid decomposition of the \(\hat \beta - \beta\) by

   # making iid decomposition of regression parameters
   betaiid <- IC(out)
   head(betaiid)
#>                x
#> [1,] -0.34616008
#> [2,] -1.44918926
#> [3,] -0.03898156
#> [4,]  0.42156050
#> [5,]  0.34253904
#> [6,] -0.07706668
   # robust standard errors
   crossprod(betaiid/NROW(betaiid))^.5
#>            x
#> x 0.02817714
   # same as 

We now look at the plot with robust standard errors

  bplot(rob,se=TRUE,robust=TRUE,col=3)

We can also make survival prediction with robust standard errors using the phreg.

  pp <-  predict(out,data[1:20,],se=TRUE,robust=TRUE)
  plot(pp,se=TRUE,whichx=1:10)

Finally, just to check that we can recover the model we also estimate the dependence parameter

tt <- twostageMLE(out,data=data)
summary(tt)
#> Dependence parameter for Clayton-Oakes model
#> Variance of Gamma distributed random effects
#> $estimates
#>                 Coef.         SE        z P-val Kendall tau        SE
#> dependence1 0.5316753 0.03497789 15.20032     0   0.2100093 0.0109146
#> 
#> $type
#> NULL
#> 
#> attr(,"class")
#> [1] "summary.mets.twostage"

Goodness of fit

The observed score process is given by \[\begin{align} U(t,\hat \beta) & = \sum_k \sum_i \int_0^t (Z_{ki} - \hat E_g ) d \hat M_{ki}^g \end{align}\] where \(g\) is strata \(g(k,i)\). The observed score has the iid decomposition \[\begin{align} \hat U(t) = \sum_k \sum_i \int_0^t (Z_{ki} - E_g) dM_{ki}^g - I(t) \sum_k \beta_k \end{align}\] where \(\beta_k\) is the iid decomposition of the score process for the true \(\beta\) \[\begin{align} \beta_k & = I(\tau)^{-1} \sum_i \int_0^\tau (Z_{ki} - E_g ) d M_{ki}^g \end{align}\] and \(I(t)\) is the derivative of the total score, \(\hat U(t,\beta))\), with respect to \(\beta\) evaluated at time \(t\).

This observed score can be resampled given it is on iid form in terms of clusters.

Now using the cumulative score process for checking proportional hazards

gout <- gof(out)
gout
#> Cumulative score process test for Proportionality:
#>   Sup|U(t)|  pval
#> x  30.24353 0.401

The p-value reflects wheter the observed score process is consistent with the model.

  plot(gout)

Computational aspects

The score processes can be resampled as in Lin, Wei, Ying (1993) using the martingale structure, such that the observed score process is resampled by \[\begin{align} \sum_k \sum_i \int_0^t g_{ki} (Z_{ki} - E_g) dN_{ki} - I(t) I^{-1}(\tau) g_{ki} \int_0^{\tau} (Z_{ki} - E_g) dN_{ki} . \end{align}\] where \(g_{ki}\) are i.i.d. standard normals.

Based on the zero mean processes we more generally with clusters can resample the score process. For resampling of score process we need \[\begin{align} U(t,\beta) & = \sum_k \sum_i g_k \int_0^t (Z_{ki} - E_g ) dM_{ki}^g \end{align}\] where \(g\) is strata. We write \(g_k\) as \(g_{ki}\) and thus repeating \(g_k\) within each cluster.

Computations are done using that \[\begin{align*} \int_0^t (Z_{ki} - E_{g}) dM_{ki}^g & = \int_0^t (Z_{ki} - E_{g}) dN_{ki}^g - \int_0^{t} (Z_{ki} - E_{g}) Y_{ki}(u) d\Lambda^g(u) \end{align*}\] therefore and summing the compensator part with the \(g_{ki}\) multipliers then gives for each strata \(g\) \[\begin{align*} & \int_0^t \frac{S_{1g}^w(u)}{S_{0g}(u)} dN_{g.}(v) - \int_0^t E_{g}(u) \frac{S_{0g}^w(u)}{S_{0g}(u)} dN_{g.}(v) \end{align*}\] with
\[\begin{align*} S_{jg}^w(t) & = \sum_{ki \in g} \exp(Z_{ki} \beta) Z_{ki}^j Y_{ki}(t) g_{ki} \\ S_{jg}(t) & = \sum_{ki \in g} \exp(Z_{ki} \beta) Z_{ki}^j Y_{ki}(t). \end{align*}\]

Cluster stratified Cox models

For clustered data it is possible to estimate the regression coefficient within clusters by using Cox’s partial likelihood stratified on clusters.

Note, here that the data is generated with a different subject specific structure, so we will not recover the \(\beta\) at 0.3 and the model will not be a proportional Cox model, we we would also expect to reject “proportionality” with the gof-test.

The model can be thought of as \[ \lambda_{g(k,i)} (t) \exp( X_{ki}^T \beta) \] where \(\lambda_g(t)\) is some cluster specific baseline.

The regression coefficient \(\beta\) can be estimated by using the partial likelihood for clusters.

 out <- phreg(Surv(time,status)~x+strata(cluster),data=data)
 summary(out)
#> 
#>     n events
#>  5000   4854
#> coeffients:
#>   Estimate     S.E.  dU^-1/2 P-value
#> x 0.406307 0.032925 0.039226       0
#> 
#> exp(coeffients):
#>   Estimate   2.5%  97.5%
#> x   1.5013 1.4074 1.6013

SessionInfo

sessionInfo()
#> R version 4.3.2 (2023-10-31)
#> Platform: aarch64-apple-darwin22.6.0 (64-bit)
#> Running under: macOS Sonoma 14.3.1
#> 
#> Matrix products: default
#> BLAS:   /Users/kkzh/.asdf/installs/R/4.3.2/lib/R/lib/libRblas.dylib 
#> LAPACK: /Users/kkzh/.asdf/installs/R/4.3.2/lib/R/lib/libRlapack.dylib;  LAPACK version 3.11.0
#> 
#> locale:
#> [1] en_US.UTF-8/en_US.UTF-8/en_US.UTF-8/C/en_US.UTF-8/en_US.UTF-8
#> 
#> time zone: Europe/Copenhagen
#> tzcode source: internal
#> 
#> attached base packages:
#> [1] stats     graphics  grDevices utils     datasets  methods   base     
#> 
#> other attached packages:
#> [1] mets_1.3.4     timereg_2.0.5  survival_3.5-7
#> 
#> loaded via a namespace (and not attached):
#>  [1] cli_3.6.2           knitr_1.45          rlang_1.1.3        
#>  [4] xfun_0.41           highr_0.10          jsonlite_1.8.8     
#>  [7] listenv_0.9.1       future.apply_1.11.1 lava_1.7.4         
#> [10] htmltools_0.5.6.1   sass_0.4.7          rmarkdown_2.25     
#> [13] grid_4.3.2          evaluate_0.23       jquerylib_0.1.4    
#> [16] fastmap_1.1.1       mvtnorm_1.2-4       yaml_2.3.7         
#> [19] numDeriv_2016.8-1.1 compiler_4.3.2      codetools_0.2-19   
#> [22] ucminf_1.2.0        Rcpp_1.0.12         future_1.33.1      
#> [25] lattice_0.22-5      digest_0.6.34       R6_2.5.1           
#> [28] parallelly_1.37.0   parallel_4.3.2      splines_4.3.2      
#> [31] bslib_0.5.1         Matrix_1.6-5        tools_4.3.2        
#> [34] globals_0.16.2      cachem_1.0.8

These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.