The terminology around that is confusing. On its own, a Gaussian process is a prior distribution over a function $f(x)$, $p(f \mid x)$. The most common (by far!) case you see discussed is when that function additionally has IID Gaussian noise added to it, call it $y$. $y = f + \epsilon$, where $\epsilon \sim N(0, \sigma^2)$. Bayes theorem here is
$$
p(f, \sigma^2 \mid x, y) = \frac{p(y \mid f, x, \sigma^2) p(f \mid x) p(\sigma^2)}{p(y \mid x)}
$$
The likelihood, $p(y \mid f, x, \sigma^2)$ is a product of normals, each with mean $f_i$ and variance $\sigma^2$, and $p(f \mid x)$ is a multivariate normal, the GP prior.
These distributions are conjugate, so $f$ can be integrated out analytically, producing the marginal likelihood. This marginal likelihood ends up being a multivariate normal whose covariance is, like you had, $K(x, x') + \sigma^2$.
So your understanding is partially correct. If the variance of the Gaussian likelihood goes to zero, you absolutely still use GPs. All that happens is that Bayes theorem above gets a bit simpler.
This answer may help you also.