I think it makes sense here to step back and simplify things. For the purpose of this answer we can think about this model:
Y ~ X + (X | G)
...in two scenarios: where X varies at the individual / unit level, and where X varies at the group level.
The motivation for fitting random slopes often arises out of the following. We have a study where we measures individuals, and we are interested in some fixed effect, ie slope of a variable. It could be the same variable measured over time, or it could be the response to different treatment levels of a variable, for example. If we had only one individual we would simply take measurements and think about a plot such as this:
set.seed(1)
X <- 1:20
Y <- 3 + X + rnorm(20, 0, 3)
ggplot(data.frame(Y, X), aes(y = Y, x = X)) + geom_point() + geom_smooth(method = 'lm', se = FALSE)

Our interest would then be in the slope of the fitted line, from the model:
> lm(Y ~ X) %>% coef()
(Intercept) X
3.062716 1.067789
Now, when we have multiple individuals, we don't want to fit seperate models for each individual, as discussed here: Difference between t-test on betas from individual regressions vs linear mixed modeling
So we want random intercepts, where each individual will have the same fixed effect (slope) for X, but a different intercept. Moreover, we naturally would expect each individual to have their own slope, so we want random slopes for X:
set.seed(1)
n.group <- 10
dt <- expand.grid(G = 1:n.group, X = 1:20)
dt$Y = 1
X <- model.matrix(~ X, dt)
myFormula <- "Y ~ X + (X | G)"
foo <- lFormula(eval(myFormula), dt)
Z <- t(as.matrix(foo$reTrms$Zt))
betas <- c(3, 1)
b1 <- rnorm(n.group, 0, 3) # random intercepts
b2 <- rnorm(n.group, 0, 0.5) # random slopes
b <- c(rbind(b1, b2))
dt$Y <- X %% betas + Z %% b + rnorm(nrow(dt), 1)
dt$G <- as.factor(dt$G)
ggplot(dt, aes(y = Y, x = X, colour = G)) + geom_point() + geom_smooth(method = 'lm', formula= y ~ x, se = FALSE)

All is good. This is a classical plot to illlustrate random slopes and intercepts. Each line represents one individual / group and has it's own intercept and slope. Note that this is not plotted from the output of a mixed model, but rather from the data itself. We fit a mixed model in order to estimate the parameters, in the case of the random effects, the variance and covariance of the random intercepts and slopes.
Now, if we let X be a group-level predictor:
dt$X <- as.numeric(dt$G) / 4
X <- model.matrix(~ X, dt)
dt$Y <- X %% betas + Z %% b + rnorm(nrow(dt), 1)
ggplot(dt, aes(y = Y, x = X, colour = G)) + geom_point() + geom_smooth(method = 'lm', formula= y ~ x, se = FALSE)

We can immediately see that each group is a vertical accumulation of points for each X value. So there is no slope for each group / individual.
This is why it does not make sense to fit random slopes for a variable that only varies at the group level. If we try to fit a model with random slopes to such data, it will almost certainly not converge, or converge to a singular fit. I say almost certainly, because as noted in the OP, we do sometimes see such model that do converge. This is why it is necessary for analysts to think about what they are doing. Plotting the data is a very good first step in many analysis tasks and can help in avoiding mistakes, and generally guide the analysis in the right direction.
ggplot(hsb10) + aes(x=size, y = math) + geom_path()+ facet_wrap(~sch.id)`
– rnorouzian Oct 07 '20 at 21:44