I have data from a live-attenuated vaccine study and I want to estimate the distribution of shedding times after vaccination. I have samples collected at multiple time points for each subject, and I can classify the results at each time point as positive or negative for shedding vaccine. Subjects are observed to either never be positive at any time point, or transition from positive to negative at later time points (or never turn negative and thus are right-censored).
IFF I could assume each subject who received the vaccine does shed for some finite amount of time, I could model this as a standard interval-censored survival analysis.
But, biologically, there is zero-inflation at the start. If one were to do the study with many more time points, one would see that some subjects never shed at all and others shed for short times that end before our first time point. These are two different states--did the vaccine "take" or not, and if so, how long was it shed for?--and should be modeled as a hurdle-survival model.
This is a pretty common situation in infectious disease challenge studies, but I am unable to locate any papers that focus on this kind of mixture survival model (the opposite of a cure mixture model), let alone an R package.
Absent this "right way" to do it, I will do something easy enough that's worked before: model the mixture parametrically using the wrong score function. I know a priori that a lognormal survival model is a good one, and so I can fit that to the proportion data at each time point, assuming a binomial scoring function, and add a free parameter that the probability of shedding at the start is between 0 and 1.
Any tips?
Thanks!