Survival analysis Part 2
Time to event models
functions
The core functions we’ll use out of the survival package include:
-
Surv()
: Creates a survival object. -
survfit()
: Fits a survival curve using either a formula, of from a previously fitted Cox model. -
coxph()
: Fits a Cox proportional hazards regression model.
Other optional functions you might use include:
-
cox.zph()
: Tests the proportional hazards assumption of a Cox regression model. -
survdiff()
: Tests for differences in survival between two groups using a log-rank / Mantel-Haenszel test.1
Surv()
creates the response variable, and typical usage takes the time to event,[^time2] and whether or not the event occured (i.e., death vs censored). survfit()
creates a survival curve that you could then display or plot. coxph()
implements the regression analysis, and models specified the same way as in regular linear models, but using the coxph()
function.
Data description
setup
take a look at the dataset
doa | dod | status | sex | dm | gcs | sbp | dbp | wbc | time2 | stroke_type | referral_from |
---|---|---|---|---|---|---|---|---|---|---|---|
17/2/2011 | 18/2/2011 | alive | male | no | 15 | 151 | 73 | 12.5 | 1 | IS | non-hospital |
20/3/2011 | 21/3/2011 | alive | male | no | 15 | 196 | 123 | 8.1 | 1 | IS | non-hospital |
9/4/2011 | 10/4/2011 | dead | female | no | 11 | 126 | 78 | 15.3 | 1 | HS | hospital |
12/4/2011 | 13/4/2011 | dead | male | no | 3 | 170 | 103 | 13.9 | 1 | IS | hospital |
12/4/2011 | 13/4/2011 | alive | female | yes | 15 | 103 | 62 | 14.7 | 1 | IS | non-hospital |
4/5/2011 | 5/5/2011 | dead | female | no | 3 | 91 | 55 | 14.2 | 1 | HS | hospital |
Creating a Survival Object
The survival object is created by the function survival::Surv
, which typically requires two arguments: event
and time
. The survival object will be used as the outcome by the survival analysis methods we explore.
Some key components of this survfit
object that will be used to create survival curves include:
-
time
: the timepoints at which the curve has a step, i.e. at least one event occurred -
surv
: the estimate of survival at the correspondingtime
The Kaplan-Meier (KM) Estimator is a non-parametric method that estimates the survival probabilities at each time an event occurs. We will use the function survival::survfit()
, which uses the KM Estimator, to estimate survival probabilities from our data.
survfit
requires two arguments:
- A formula object where the outcome is a survival object
- A data frame
km_result <- survival::survfit(
surv_object ~ 1,
data = surv_data
)
The Output of survfit
#> # A tibble: 17 × 5
#> TIME N_RISK N_EVENT CENSOR SURVIVAL
#> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 1 213 9 14 0.958
#> 2 2 190 4 20 0.938
#> 3 3 166 4 32 0.915
#> 4 4 130 4 36 0.887
#> 5 5 90 5 20 0.838
#> 6 6 65 3 6 0.799
#> 7 7 56 4 6 0.742
#> 8 9 42 1 8 0.724
#> 9 10 37 1 1 0.705
#> 10 12 33 4 5 0.619
#> 11 14 24 2 2 0.568
#> 12 18 19 1 3 0.538
#> 13 22 15 1 5 0.502
#> 14 25 9 2 4 0.390
#> 15 28 5 1 1 0.312
#> 16 29 4 1 0 0.234
#> 17 41 2 1 1 0.117
The survival probabilities for all patients:
>KM <- survfit(Surv(time = time2,event = status == "dead" ) ~ 1,
data = surv_data)
>summary(KM)
: survfit(formula = Surv(time = time2, event = status == "dead") ~
Call1, data = surv_data)
95% CI upper 95% CI
time n.risk n.event survival std.err lower 1 213 9 0.958 0.0138 0.9311 0.985
2 190 4 0.938 0.0168 0.9053 0.971
3 166 4 0.915 0.0198 0.8770 0.955
4 130 4 0.887 0.0237 0.8416 0.934
5 90 5 0.838 0.0310 0.7790 0.901
6 65 3 0.799 0.0367 0.7301 0.874
7 56 4 0.742 0.0438 0.6608 0.833
9 42 1 0.724 0.0462 0.6391 0.821
10 37 1 0.705 0.0489 0.6150 0.807
12 33 4 0.619 0.0587 0.5142 0.746
14 24 2 0.568 0.0642 0.4548 0.708
18 19 1 0.538 0.0674 0.4206 0.687
22 15 1 0.502 0.0718 0.3792 0.664
25 9 2 0.390 0.0892 0.2494 0.611
28 5 1 0.312 0.0998 0.1669 0.584
29 4 1 0.234 0.1009 0.1007 0.545
41 2 1 0.117 0.0970 0.0231 0.593
Next, we will estimate the survival probabilities for stroke type:
- this will give us two tables ,relating to two factors that we have in the data
>KM_str_type2 <- survfit(Surv(time = time2,
event = status == "dead" ) ~ stroke_type,
data = surv_data)
>summary(KM_str_type2)
: survfit(formula = Surv(time = time2, event = status == "dead") ~
Calldata = surv_data)
stroke_type,
=HS
stroke_type95% CI upper 95% CI
time n.risk n.event survival std.err lower 1 69 6 0.913 0.0339 0.8489 0.982
2 61 1 0.898 0.0365 0.8293 0.973
3 58 4 0.836 0.0453 0.7520 0.930
4 52 2 0.804 0.0489 0.7136 0.906
5 47 4 0.736 0.0554 0.6346 0.853
6 38 2 0.697 0.0589 0.5905 0.822
7 34 2 0.656 0.0621 0.5447 0.790
9 30 1 0.634 0.0638 0.5205 0.772
10 27 1 0.611 0.0656 0.4945 0.754
12 24 2 0.560 0.0693 0.4390 0.713
14 19 1 0.530 0.0717 0.4068 0.691
18 15 1 0.495 0.0751 0.3675 0.666
22 11 1 0.450 0.0806 0.3166 0.639
25 6 2 0.300 0.1019 0.1541 0.584
29 2 1 0.150 0.1176 0.0322 0.698
{{% callout warning %}} + this is another output ,technically they should be together in one call :::
=IS
stroke_type95% CI upper 95% CI
time n.risk n.event survival std.err lower 1 144 3 0.979 0.0119 0.956 1.000
2 129 3 0.956 0.0174 0.923 0.991
4 78 2 0.932 0.0241 0.886 0.980
5 43 1 0.910 0.0318 0.850 0.975
6 27 1 0.876 0.0451 0.792 0.970
7 22 2 0.797 0.0676 0.675 0.941
12 9 2 0.620 0.1223 0.421 0.912
14 5 1 0.496 0.1479 0.276 0.890
28 3 1 0.331 0.1671 0.123 0.890
41 1 1 0.000 NaN NA NA
The KM estimate provides the survival probabilities. We can plot these probabilities to look at the trend of survival over time. The plot provides
- survival probability on the \(y-axis\)
- time on the \(x-axis\)
Plotting the Survival Function
survminer::ggsurvplot(
km_result,
pval = TRUE,
conf.int = TRUE,
xlab = "Time",
ylab = "Probability of dying"
)
Using the KM Estimator to Plot Multiple Survival Functions
km_result_jobs <- survival::survfit(surv_object ~stroke_type, data = surv_data)
survminer::ggsurvplot(km_result_jobs, pval = TRUE, xlab = "Time", ylab = "Probability of dying")
Making it more fency
- i always love to make the graphs more fency
- i have used ggplot2 extended functionalities to make the plots pretty neat
library(paletteer)
p1<-ggsurvplot(KM_str_type2,
data = surv_data,
palette = paletteer_d("ggsci::light_blue_material")[seq(2,10,2)],
size = 1.2, conf.int = FALSE,
legend.labs = levels(surv_data$stroke_type),
legend.title = "",
ggtheme = theme_minimal() +
theme(plot.title = element_text(face = "bold")),
title = "Probability of dying",
xlab = "Time",
ylab = "Probability of dying",
legend = "bottom", censor = FALSE)
p1$plot +
scale_x_continuous(expand = c(0, 0), breaks = seq(1,43,1),
labels = seq(1,43,1),
limits = c(0, 820)) +
scale_y_continuous(expand = c(0, 0), breaks = seq(0,1,0.2),
labels = scales::percent_format(accuracy = 1)) +
theme_classic()
We can perform the Kaplan-Meier estimates for variable dm too:
KM_dm <- survfit(Surv(time = time2,
event = status == "dead" ) ~ dm,
data = surv_data)
summary(KM_dm)
#> Call: survfit(formula = Surv(time = time2, event = status == "dead") ~
#> dm, data = surv_data)
#>
#> dm=no
#> time n.risk n.event survival std.err lower 95% CI upper 95% CI
#> 1 141 8 0.943 0.0195 0.9058 0.982
#> 2 122 4 0.912 0.0242 0.8661 0.961
#> 3 102 2 0.894 0.0268 0.8434 0.949
#> 4 82 2 0.873 0.0303 0.8152 0.934
#> 5 54 5 0.792 0.0441 0.7100 0.883
#> 6 40 3 0.732 0.0524 0.6366 0.843
#> 7 34 2 0.689 0.0575 0.5854 0.812
#> 10 24 1 0.661 0.0619 0.5498 0.794
#> 12 20 4 0.529 0.0771 0.3971 0.703
#> 18 13 1 0.488 0.0812 0.3521 0.676
#> 22 9 1 0.434 0.0884 0.2908 0.647
#> 25 4 1 0.325 0.1149 0.1627 0.650
#> 29 3 1 0.217 0.1171 0.0752 0.625
#>
#> dm=yes
#> time n.risk n.event survival std.err lower 95% CI upper 95% CI
#> 1 72 1 0.986 0.0138 0.9594 1.000
#> 3 64 2 0.955 0.0253 0.9070 1.000
#> 4 48 2 0.915 0.0367 0.8463 0.990
#> 7 22 2 0.832 0.0653 0.7137 0.971
#> 9 15 1 0.777 0.0811 0.6330 0.953
#> 14 9 2 0.604 0.1248 0.4030 0.906
#> 25 5 1 0.483 0.1471 0.2662 0.878
#> 28 2 1 0.242 0.1860 0.0534 1.000
#> 41 1 1 0.000 NaN NA NA
And then we can plot the survival estimates for patients with and without diabetes:
p2<-ggsurvplot(KM_dm,
data = surv_data,
palette = paletteer_c("grDevices::Set 2", 12),
size = 1.2, conf.int = FALSE,
legend.labs = levels(surv_data$dm),
legend.title = "",
ggtheme = theme_minimal() +
theme(plot.title = element_text(face = "bold")),
title = "Probability of dying",
xlab = "Time till discharge",
ylab = "Probability of dying",
legend = "bottom", censor = FALSE)
p2
- Typically we will also want to see the numbers at risk in a table below the x-axis. We can add this using
add_risktable()
:
library(ggsurvfit)
survfit2(Surv(time = time2,event = status == "dead" ) ~ dm, data = surv_data) %>%
ggsurvfit() +
labs(
x = "Days",
y = "Overall survival probability"
) +
add_confidence_interval() +
add_risktable()
There are a number of available tests to compare the survival estimates between groups based on KM. The tests include:
- log-rank (default)
- peto-peto test
Log-rank test
to answer question if the survival estimates are different between levels or groups we can use statistical tests for example the log rank and the peto-peto tests.
For all the test, the null hypothesis is that that the survival estimates between levels or groups are not different. For example, to do that:
>survdiff(Surv(time = time2,
event = status == "dead") ~ stroke_type,
data = surv_data,
rho = 0)
:
Callsurvdiff(formula = Surv(time = time2, event = status == "dead") ~
data = surv_data, rho = 0)
stroke_type,
Expected (O-E)^2/E (O-E)^2/V
N Observed =HS 69 31 24.2 1.92 4.51
stroke_type=IS 144 17 23.8 1.95 4.51
stroke_type
= 4.5 on 1 degrees of freedom, p= 0.03 Chisq
The survival estimates between the surv_data types (IS vs HS groups) are different at the level of \(5\%\) significance (p-value = 0.03).
And for the survival estimates based on diabetes status:
>survdiff(Surv(time = time2,
event = status == "dead") ~ dm,
data = surv_data,
rho = 0)
:
Callsurvdiff(formula = Surv(time = time2, event = status == "dead") ~
data = surv_data, rho = 0)
dm,
Expected (O-E)^2/E (O-E)^2/V
N Observed =no 141 35 29.8 0.919 2.54
dm=yes 72 13 18.2 1.500 2.54
dm
= 2.5 on 1 degrees of freedom, p= 0.1 Chisq
The survival estimates between patients with and without diabetes (dm status yes vs no groups) are not different (p-value = 0.1).
Cox PH General model
The Cox model is expressed by the hazard function denoted by \(h(t)\). This model can be used to fit univariable and multivariable regression models that have survival outcomes. The hazard function can be interpreted as the risk of dying at time t. It can be estimated as follow:
\(h(t, X_{i}) = h_{0}(t)e^{ \sum_{j=1}^{p} \beta_{j}X_{i,j}} = h_{0}(t)exp(\beta_{1}X_{i,1} + ... +\ beta_{p}X_{i,p})\)
where:
- \(h(t)\) is the hazard, the instantaneous rate at which events occur.
- \(h_{0}(t)\) is called the baseline hazards (when all X’s are equal to 0), depends on \(t\)
- \(X = (X_{1}, X_{2},..., X_{p})\) explanatory/predictor variables
- \(e^{ \sum_{i=1}^{p} \beta_{i}X_{i}}\), depends only on X’s, called .
Because the baseline hazard \(h_{0}(t)\) is an unspecified function, the Cox model us a semiparametric model.
Advantages of the model: “robust” model, so that the results from using the Cox model will closely approximate the results for the correct parametric model.
The Cox model can be written as a multiple linear regression of the logarithm of the hazard on the variables \(X_{i}\), with the baseline hazard, \(h_{0}(t)\), being an ‘intercept’ term that varies with time.
\[log(h(t, X_{i})) = log(h_{0}(t)) + \sum_{j=1}^{p} \beta_{j}X_{i,j}\]
We can compute the hazard ratio, which is the ratio of hazards between two groups at any particular point in time: “hazard for one individual divided by the hazard for a different individual”.
\[\hat{HR} = \frac{\hat{h}(t, X^{*})}{\hat{h}(t, X)} = e^{ \sum_{i=1}^{p} \beta_{i} (X^{*}_{i} - X_{i})}\]
with:
\(X^{*}\): set of predictors for one individual
X: set of predictors for the other individual
This model shows that the hazard ratio is equal to \(e^{ \sum_{i=1}^{p} \beta_{i} (X^{*}_{i} - X_{i})}\), and remains constant over time t (hence the name proportional hazards regression). In this sense, we do not need the baseline hazard because we can interpret coefficients as hazard ratios.
A hazard ratio above 1 indicates a covariate that is positively associated with the event probability, and thus negatively associated with the length of survival.
In summary,
- HR = 1 : No effect
- HR < 1: Reduction in the hazard
- HR > 1: Increase in Hazard
As a note, in clinical studies, a covariate with hazard ratio :
- greater than 1 (i.e.: b>0) is called bad prognostic factor.
- smaller than 1 (i.e.: b<0) is called good prognostic factor.
As a consequence, a major assumption of this model is that the HR is constant over time because it is independent of time. Or equivalently that the hazard for one individual is proportional to the hazard for any other individual, where the proportionality constant is independent of time.
It is possible, nevertheless, to consider X’s which do involve t. Such X’s are called time-dependent variables. If time-dependent variables are considered, the Cox model form may still be used, but such a model no longer satisfies the PH assumption, and is called the extended Cox model.
- The coxph() function uses the same syntax as lm(), glm(), etc. The response variable you create with Surv() goes on the left hand side of the formula, specified with a ~. Explanatory variables go on the right side.
COX PH model with surv_data type variable only
(surv_data_stype <- coxph(Surv(time2,status == 'dead') ~ stroke_type
,data = surv_data))
#> Call:
#> coxph(formula = Surv(time2, status == "dead") ~ stroke_type,
#> data = surv_data)
#>
#> coef exp(coef) se(coef) z p
#> stroke_typeIS -0.6622 0.5157 0.3172 -2.088 0.0368
#>
#> Likelihood ratio test=4.52 on 1 df, p=0.03344
#> n= 213, number of events= 48
The effect of surv_data type is significantly related to survival (p-value = 0.0368), with better survival in Ischaemic surv_data in comparison to the other type (hazard ratio of dying = 0.5157).
The model is statistically significant. That 0.03344 p-value of the model is really close to the p = 0.03 p-value we saw on the Kaplan-Meier nodel as well as the likelihood ratio test = 4.52 is close to the log-rank chi-square (4.5) in the Kaplan-Meier model.
\(e^{\beta_{1}}\) = \(e^{-0.6622}\) = 0.5157 is the hazard ratio - the multiplicative effect of that variable on the hazard rate (for each unit increase in that variable). Ischaemic surv_data patients have 0.588 (~ 60%) times the hazard of dying in comparison to haemorage.
this is important when perfoming statistical analysis
- first build a model with all varibles
>surv_data_stype <- coxph(Surv(time2,status == 'dead') ~ gcs + stroke_type
+ sex + dm + sbp ,data = surv_data)
>summary(surv_data_stype)
:
Callcoxph(formula = Surv(time2, status == "dead") ~ gcs + stroke_type +
+ dm + sbp, data = surv_data)
sex
= 213, number of events= 48
n
exp(coef) se(coef) z Pr(>|z|)
coef -0.170038 0.843633 0.037167 -4.575 4.76e-06 ***
gcs -0.103523 0.901655 0.346749 -0.299 0.765
stroke_typeIS -0.203488 0.815880 0.334159 -0.609 0.543
sexmale -0.439913 0.644093 0.343960 -1.279 0.201
dmyes -0.001765 0.998237 0.004017 -0.439 0.660
sbp ---
: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 Signif. codes
exp(coef) exp(-coef) lower .95 upper .95
0.8436 1.185 0.7844 0.9074
gcs 0.9017 1.109 0.4570 1.7791
stroke_typeIS 0.8159 1.226 0.4238 1.5706
sexmale 0.6441 1.553 0.3282 1.2639
dmyes 0.9982 1.002 0.9904 1.0061
sbp
= 0.78 (se = 0.035 )
Concordance= 28.88 on 5 df, p=2e-05
Likelihood ratio test= 27.71 on 5 df, p=4e-05
Wald test Score (logrank) test = 31.45 on 5 df, p=8e-06
- the estimate which is the log hazard. If you exponentiate it, you will get hazard ratio
- the standard error
- the p-value
- the confidence intervals for the log hazard
using rms package
variable importance
now we create univariate models as in logistic regression
By using tbl_uvregression()
we can generate simple univariable model for all covariates in one line of code. In return, we get the crude HR for all the covariates of interest.
library(gt)
library(gtsummary)
surv_data |>
dplyr::select(time2, status, sex, dm, gcs, sbp, dbp, wbc,
stroke_type) |>
tbl_uvregression(
method = coxph,
y = Surv(time2, event = status == 'dead'),
exponentiate = TRUE,
pvalue_fun = ~style_pvalue(.x, digits = 3)
) |>
as_gt()
Characteristic | N | HR1 | 95% CI1 | p-value |
---|---|---|---|---|
sex | 213 | |||
female | — | — | ||
male | 0.71 | 0.37, 1.36 | 0.299 | |
dm | 213 | |||
no | — | — | ||
yes | 0.60 | 0.31, 1.13 | 0.112 | |
gcs | 213 | 0.84 | 0.79, 0.90 | <0.001 |
sbp | 213 | 1.00 | 0.99, 1.01 | 0.617 |
dbp | 213 | 1.00 | 0.98, 1.01 | 0.772 |
wbc | 213 | 1.04 | 0.97, 1.11 | 0.270 |
stroke_type | 213 | |||
HS | — | — | ||
IS | 0.52 | 0.28, 0.96 | 0.037 | |
1 HR = Hazard Ratio, CI = Confidence Interval |
- it is clear that
gcs
has the least p-value hence can be included in the model first
surv_data_gcs <- coxph(Surv(time = time2,event = status == 'dead') ~ gcs,
data = surv_data)
summary(surv_data_gcs)
#> Call:
#> coxph(formula = Surv(time = time2, event = status == "dead") ~
#> gcs, data = surv_data)
#>
#> n= 213, number of events= 48
#>
#> coef exp(coef) se(coef) z Pr(>|z|)
#> gcs -0.17454 0.83984 0.03431 -5.087 3.63e-07 ***
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> exp(coef) exp(-coef) lower .95 upper .95
#> gcs 0.8398 1.191 0.7852 0.8983
#>
#> Concordance= 0.763 (se = 0.039 )
#> Likelihood ratio test= 26.01 on 1 df, p=3e-07
#> Wald test = 25.88 on 1 df, p=4e-07
#> Score (logrank) test = 29.33 on 1 df, p=6e-08
The simple Cox PH model with covariate gcs shows that with each one unit increase in gcs, the crude log hazard for death changes by a factor of \(-0.175\).
lets tidy it up
tidy(surv_data_gcs,
exponentiate = TRUE,
conf.int = TRUE)
#> # A tibble: 1 × 7
#> term estimate std.error statistic p.value conf.low conf.high
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 gcs 0.840 0.0343 -5.09 0.000000363 0.785 0.898
When we exponentiate the log HR, the simple Cox PH model shows that with each one unit increase in gcs, the crude risk for death decreases for about \(16\%\) and the of decrease are between \(95\% CI (0.785, 0.898)\). The relationship between surv_data death and gcs is highly significant (p-value \(< 0.0001\)) when not adjusting for other covariates.
next up we add surv_data type
surv_data_mv <-
coxph(Surv(time = time2,
event = status == 'dead') ~ stroke_type + gcs ,
data = surv_data)
tidy(surv_data_mv, exponentiate = TRUE, conf.int = TRUE)
#> # A tibble: 2 × 7
#> term estimate std.error statistic p.value conf.low conf.high
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 stroke_typeIS 0.774 0.323 -0.794 0.427 0.411 1.46
#> 2 gcs 0.846 0.0357 -4.68 0.00000290 0.789 0.908
lets test if surv_data type is worth it
anova(surv_data_mv,surv_data_gcs,test="LRT")
#> Analysis of Deviance Table
#> Cox model: response is Surv(time = time2, event = status == "dead")
#> Model 1: ~ stroke_type + gcs
#> Model 2: ~ gcs
#> loglik Chisq Df Pr(>|Chi|)
#> 1 -187.48
#> 2 -187.80 0.6422 1 0.4229
- The variable is not significant
we try another one
surv_data_dm <-
coxph(Surv(time = time2,
event = status == 'dead') ~ dm + gcs ,
data = surv_data)
tidy(surv_data_dm, exponentiate = TRUE, conf.int = TRUE)
#> # A tibble: 2 × 7
#> term estimate std.error statistic p.value conf.low conf.high
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 dmyes 0.625 0.328 -1.43 0.152 0.329 1.19
#> 2 gcs 0.840 0.0347 -5.04 0.000000466 0.785 0.899
lets test if dm is worth it
anova(surv_data_dm,surv_data_gcs,test="LRT")
#> Analysis of Deviance Table
#> Cox model: response is Surv(time = time2, event = status == "dead")
#> Model 1: ~ dm + gcs
#> Model 2: ~ gcs
#> loglik Chisq Df Pr(>|Chi|)
#> 1 -186.71
#> 2 -187.80 2.181 1 0.1397
- still not worth is it , we can go on and on…
lets do a backward elimination
fatal_mv1<-rms::cph(Surv(time2,status == 'dead') ~ gcs + stroke_type+dbp+wbc
+ sex + dm + sbp ,data = surv_data)
fastbw(fatal_mv1)
#>
#> Deleted Chi-Sq d.f. P Residual d.f. P AIC
#> wbc 0.07 1 0.7881 0.07 1 0.7881 -1.93
#> stroke_type 0.14 1 0.7130 0.21 2 0.9015 -3.79
#> sex 0.16 1 0.6934 0.36 3 0.9478 -5.64
#> sbp 0.39 1 0.5326 0.75 4 0.9447 -7.25
#> dbp 1.12 1 0.2897 1.87 5 0.8664 -8.13
#> dm 2.03 1 0.1542 3.90 6 0.6898 -8.10
#>
#> Approximate Estimates after Deleting Factors
#>
#> Coef S.E. Wald Z P
#> gcs -0.1762 0.03495 -5.041 4.635e-07
#>
#> Factors in Final Model
#>
#> [1] gcs
- final model has gcs variable
The Cox proportional hazards model makes several assumptions. We use residuals methods to:
- check the proportional hazards assumption with the Schoenfeld residuals
- detect nonlinearity in relationship between the log hazard and the covariates using Martingale residual
- examining influential observations (or outliers) with deviance residual (symmetric transformation of the martinguale residuals), to examine influential observations
Testing proportional hazard
The proportional hazard assumption is supported by a non-significant relationship between residuals and time, and refuted by a significant relationship.
We can test with the Goodness of Fit (GOF) approach based on the residuals defined by Schoenfeld.
The idea behind the statistical test is that if the PH assumption holds for a particular covariate then the Schoenfeld residuals for that covariate will not be related to survival time.
For each covariate, the function cox.zph() correlates the corresponding set of scaled Schoenfeld residuals with time, to test for independence between residuals and time. Additionally, it performs a global test for the model as a whole.
ggcoxzph(test.ph)
From the output above, the test is not statistically significant, and therefore the global test is also not statistically significant. Therefore, we can assume the proportional hazards.
In the graphical diagnostic using the function ggcoxzph() [in the survminer package], the solid line is a smoothing spline fit to the plot, with the dashed lines representing a +/- 2-standard-error band around the fit. From the graphical inspection, there is no pattern with time. The assumption of proportional hazards appears to be supported for the covariates sex (which is, recall, a two-level factor, accounting for the two bands in the graph).
another approach
Another approach is to graphically check the PH assumption by comparing -log–log survival curves. A log–log survival curve is simply a transformation of an estimated survival curve that results from taking the natural log of an estimated survival probability twice.
\(h(t, X) = h_{0}(t)e^{ \sum_{j=1}^{p} \beta_{j}X_{j}}\) which is equivalent to \(S(t,X) = [S_{0}(t)]^{e^{ \sum_{j=1}^{p} \beta_{j}X_{j}}}\)
Therefore, \[-ln(-ln(S(t, X)))\]
\[= -ln(-ln([S_{0}(t)]^{e^{ \sum_{j=1}^{p} \beta_{j}X_{j}}}))\] \[= -ln[-ln(S_{0}(t))] - ln[e^{ \sum_{j=1}^{p} \beta_{j}X_{j}}]\]
\[= - \sum_{j=1}^{p} \beta_{j} X_{j} - ln[-ln(S_{0}(t))]\]
Therefore, the corresponding log–log curves for these individuals are given as shown here, where we have simply substituted \(X_1\) and \(X_2\) for \(X\) in the previous expression for the log–log curve for any individual \(X\).
\[ln(-ln(S(t, X_1))) - ln(-ln(S(t, X_2)))\] \[= \sum_{j=1}^{p} \beta_{j} (X_{1j} - X_{2j})\]
The baseline survival function has dropped out, so that the difference in log–log curves involves an expression that does not involve time t. The above formula says that if we use a Cox PH model and we plot the estimated -log–log survival curves for two groups of individuals on the same graph, the two plots would be approximately parallel.
The distance between the two curves is the linear expression involving the differences in predictor values, which does not involve time. Note, in general, if the vertical distance between two curves is constant, then the curves are parallel.
m = survfit(Surv(time2,status == 'dead') ~ stroke_type
,data = surv_data)
s = summary(m)
s_table = data.frame(s$strata, s$time, s$n.risk, s$n.event, s$n.censor, s$surv, s$lower, s$upper)
s_table = s_table %>%
rename(strata=s.strata, time=s.time, surv=s.surv, lower=s.lower, upper=s.upper) %>%
mutate(negloglogsurv=-log(-log(surv)))
plot<-ggplot(s_table, aes(x=time, y=negloglogsurv, color=strata)) +
geom_line(size=1.25) +
theme(text=element_text(size=16),
plot.title=element_text(hjust=0.5)) +
ggthemes::scale_colour_tableau()+
tvthemes::theme_avatar()+
ylab("-ln(-ln(S(t)))") +
ggtitle("-log-log plot of survival time by surv_data type")
plot
The two curve cross, therefore this result suggests that the two groups in surv_data type do not satisfy the PH assumption.
To test influential observations or outliers, we can visualize either: the dfbeta values or the deviance residuals.
- dfbeta values
This plot produces the estimated changes in the coefficients divided by their standard errors. The comparison of the magnitudes of the largest dfbeta values to the regression coefficients suggests that none of the observations is terribly influential individually.
ggcoxdiagnostics(surv_data_stype, type = "dfbeta",
linear.predictions = FALSE, ggtheme = theme_bw())
{{% callout note%}}
2) deviance residuals
The deviance residual is a normalized transformation of the martingale residual. These residuals should be roughly symmetrically distributed about zero with a standard deviation of 1.
:::
ggcoxdiagnostics(stroke_type, type = "deviance",
linear.predictions = FALSE, tvtheme = theme_avatar())
ggcoxdiagnostics(stroke_type, type = "martingale",
linear.predictions = FALSE, ggtheme = theme_bw())
ggcoxdiagnostics(stroke_type, type = "deviance", ox.scale = 'time')
ggcoxdiagnostics(stroke_type, type = "deviance", ox.scale = "linear.predictions")
ggcoxdiagnostics(stroke_type, type = "schoenfeld", ox.scale = "observation.id")
Footnotes
Cox regression and the logrank test from
survdiff
are going to give you similar results most of the time. The log-rank test is asking if survival curves differ significantly between two groups. Cox regression is asking which of many categorical or continuous variables significantly affect survival.↩︎