I’m newly involved in a project, where we have access to individual patient recruitment data of 300 trials (>100k patients) + info on these trials. As most will know, most trials struggle with recruitment so there’s a lot of interest in being able to predict recruitment duration.
There’s been some work on this, e.g., here, here, here, and here. The approaches seem suboptimal, e.g., assume constant rate of accrual, or discard a lot of information, e.g., reduce data to quaretly recruitment numbers.
The main question that interests us is, after how much information, e.g., 10%, 20%, …, 50% the prediction of the total recruitment time does not improve meaningfully. In other words, if a trial has recruited 10% of its target sample size, and one uses this information make a prediction about how long the total recruitment is going to be, will this already be reasonably accurate or does one have to wait until e.g., 50%. I’m also aware that it might be “easy” to find such a threshold, even though none exists, and would like to avoid that.
Intuitively, I would have approach this as follows:
- Have the data in long format, e.g., weekly data, where y is number of participants recruited in that discrete time interval for that trial.
- Fit a flexible cumulative logit model, e.g.,:
y ~ f(time) + n_centers + … + (1|trial_id)
- Use this model to predict expected number of patients recruited for a large time grid to create a curve → compare against the actually observed recruitment curve.
- serial correlation / markov structure might also be interesting to explore
We could fit the same model (except maybe for how time is modelled), once on 10% of the data, once on 20% of the data, etc.
I’ve never seen someone do something similar, so there are surely many issues I’m missing + this all seems rather unelegant…
I’m unsure how one should evaluate and compare the predictions from these models, especially as the underlying data are not the same. And of course the model prediction will be better with more data, so I’m not even sure how one would go about saying that they are not meaningfully better anymore.
Of course, there are other various issues. E.g., trials might have taken measures because of slow recruitment, i.e., opened new centers, which complicates things further.
Would love to get some feedback on my thoughts. Thanks!