It depends on how this log-likelihood value is calculated. When you assume a Gaussian error distribution, maximizing log-likelihood is equivalent to minimizing the sum of squared residuals. Since maximization does not depend on constants out front, it is not clear to me how exactly this log-likelihood is calculated. Consquently, I would not use that value. I would directly calculate the sum of the squared residuals: $\overset{N}{\underset{i=1}{\sum}}\left(
y_i-\hat y_i
\right)^2$. If you want to normalize this to get some kind of $R^2$, you can calculate $R^2=1-\left(\dfrac{
\overset{N}{\underset{i=1}{\sum}}\left(
y_i-\hat y_i
\right)^2
}{
\overset{N}{\underset{i=1}{\sum}}\left(
y_i-\bar y
\right)^2
}\right)$.
If you apply this formula to your linear model, you will get the same value of $R^2$ are your software returns (unless you lack an intercept in the linear model). Likewise, you can use this equation to convert your model $R^2$ to a sum of squared residuals by calculating the denominator term (though I would expect it to be easier just to calculate the sum of squared residuals).
As far as determining if there is a statistical difference between the two models, first, consider what exactly you mean. Especially if sample sizes are large, statistics can detect very small differences that might not be of interest. (My comments here about the Princess and the Pea fairy tale concern such a situation.) However, if you want to calculate some statistics about differences in model performance, Benavoli et al. (2017) give some standard ways of doing such a comparison and also make a strong argument for why their proposed approach is superior. Even if you do not buy their argument, the paper at least goes through more standard approaches.
Benavoli et al. (2017) deal with classification accuracy, rather than regression metrics, but I see no reason why their proposed or referenced approachs could not apply to $R^2$ or the sum of squared residuals.
REFERENCE
Benavoli, Alessio, et al. "Time for a change: a tutorial for comparing multiple classifiers through Bayesian analysis." The Journal of Machine Learning Research 18.1 (2017): 2653-2688.