There are a bunch of factors that could go into this. Some cases where a $p$ value can be non-informative...
- Poorly captured functions. For example, if we fit some nonlinear data with a linear regression and the predictors are deemed statistically significant, then the relationship captured means a lot less than it is presented as. This can sometimes be completely caused by error (the function is very noisy), or be completely independent of it (the regression simply draws a line through data in just the right way and erroneously speaks to the relationship).
- Sample size and effect size. If you have a billion people measured on the relationship between two variables but the effect has an extremely poor magnitude, the relationship will be "statistically significant" but not "practically significant." Here the standard error should be very low, but the $p$ value really isn't that informative. The opposite can also happen where you have a low sample size and a statistically significant effect, but this is captured purely by chance. However, you can also have the case where the effect is actually present, but with a low sample size, the $p$ value is still uninformative. So the context will change how this plays out and error can weight all of those scenarios, but the standard error is usually negatively correlated with sample size.
- Correlated errors. If your data has some correlated errors but you don't model them into the regression, then your coefficients are downward biased and are misleading. For the opposite reason, if you model in the random effects and the fixed effects become statistically non-significant, then you are capturing a more "true" model despite the $p$ values being high. Both of these scenarios are by nature caused by the noise in the model.
- Unaccounted for attenuation bias. You could have a situation where the $p$ value is high for some scale made of many items used as a predictor. However, the actual relationship between the latent construct and outcome is quite strong and is only being distorted by noise from bad items, which inflates the error in the estimates. If the items were corrected and were accurately measuring the construct, then the high $p$ value could be simply attributed to attenuation bias. The opposite can also happen where you get high reliability items (low error in item reliability), but the predictor has no effect on the outcome because the relationship is poor.
- The actual effect is by nature noisy. You may run one study where the function between the variables is properly accounted for, but because it is impossible to measure with certainty, it ends up with a high $p$ value. ROC curves were originally crafted in England during WWII because of signal-to-noise concerns with shooting down German planes. Properly calibrating the data to the function may take multiple studies, of which the error could be uninformative of the influence behind the relationship.
I will state in any case that $p$ values are usually terrible metrics for how good or bad a model is, along with it's regressors. There is a lot more information one should be concerned with, and $p$ values typically distract from the actual modeling process that people should be engaging in. Standard error can at least approximate the noise in our estimate, whereas a $p$ value can completely misinterpret the relationship between variables (and often is misinterpreted).
Below is a great example, where I have simulated the same model twice, but I have tweaked the standard error to be lower in the second model. Both are completely misfitting the data, but are "statistically significant."
#### Poorly Captured Function ####
set.seed(123)
par(mfrow=c(1,2))
x <- runif(100,0,8)
y <- cos(x) + rnorm(100,sd=.5)
fit <- lm(y ~ x)
summary(fit)
plot(x,y,main="Higher Error Model")
abline(fit,col="red")
set.seed(123)
x2 <- runif(100,0,8)
y2 <- cos(x2) + rnorm(100,sd=.1)
fit2 <- lm(y2 ~ x2)
summary(fit2)
plot(x2,y2,main="Lower Error Model")
abline(fit2,col="red")

Another simulated example which shows high sample size/low error with a nearly nonexistent effect compared to a low sample size/high error with a high magnitude effect that isn't getting captured well.
#### Sample Size and Error ####
par(mfrow=c(1,2))
x3 <- rnorm(1e5)
y3 <- (x3*.01) + rnorm(1e5)
plot(x3,y3,main="High Sample Size, Low Error")
abline(fit3,col="red")
fit3 <- lm(y3 ~ x3)
summary(fit3)
x4 <- rnorm(5)
y4 <- (x4*20) + rnorm(5,sd=20)
fit4 <- lm(y4 ~ x4)
summary(fit4)
plot(x4,y4,main="Low Sample Size, High Error")
abline(fit4,col="red")

Another classic example is the Simpson Paradox. Using the iris data in R, we can see two model fits, one with low error and a statistically significant effect, another with a high error and no statistically significant effect. Both models are still wrong, and the $p$ value doesn't say enough to tell the actual story.
#### Classic Simpson Paradox ####
fit.petal <- lm(Petal.Length ~ Petal.Width,iris)
fit.sepal <- lm(Sepal.Length ~ Sepal.Width,iris)
summary(fit.petal)
summary(fit.sepal)
par(mfrow=c(1,2))
plot(iris$Petal.Width,
iris$Petal.Length,
bg=iris$Species,
pch=23,
main="Petal Regression")
abline(fit.petal,
col="red")
plot(iris$Sepal.Width,
iris$Sepal.Length,
bg=iris$Species,
pch=23,
main="Sepal Regression")
abline(fit.sepal,
col="red")
