We have developed a model for some real data and we use EM algorithm for optimization of the model (parameters). In first phase we generate synthetic data according to the model (with some known parameters) to validate that our inference code is bug free and the results are (perhaps) strange. We divide the synthetic data into train and test parts. Actually in first iterations of EM algorithm, (both) negative-log-likelihood of test data (EM objective) and parameters error (Mean Square Error, etc) decrease successfully. But in the later iterations of EM, We see that while the negative of log-likelihood decreases on test data slowly the parameters error starts to diverge. I draw a schematic figure of the situation below:

I should underline that the negative-log-likelihood of test never increases but its decreasing just slows down dramatically.
I am wondering why this situation happens? Can we consider it as over-fitting ? (and if yes, why?)