In their fantastic book Applied Longitudinal Data Analysis: Modeling Change and event Occurrence Singer and Willett advocate a iterative model comparison technique for testing the effect of predictors: starting with a baseline model and then adding predictors, one at a time, to a series of subsequent models, using a likelihood ratio test to determine whether the addition of the new predictor to the model significantly reduces the -2LL deviance. These are nested models so each LR test in effect tests the effect of each new predictor. If the -2LL deviance is significantly reduced the new model retained, if not the previous model is retained.
They say they recommend this technique over simply including all the predictors one intends to test in a single regression model and reporting the confidence intervals and p-values.
I have accepted their recommendation however I don't really understand why (i.e. on what grounds) they consider their iterative approach superior to the 'all-in' approach.
The reason I ask is that I have conducted analyses for several papers now where my co-authors have been very confused by the iterative approach. It could be I am not explaining it well enough but I would like some more solid justification for using it than 'because Singer and Willett said so'.
A plain-language explanation would be very much appreciated.