Learning about MLE involves the optimisation of log-likelihood, which allows us to get the "best" value of theta for observing a certain sample.
What then is the purpose of showing the same estimator maximises the expectation of log-likelihood? Do we not take it as given that since it maximises the log-likelihood, that it would definitely be the case that adding an expectation (mean) doesn't really value add anything?
Thank you!
