I have to admit that I disagree with Dave's answer (but I did upvote it, because it is useful). In my opinion, it makes little sense to consider your loss as a function of two variables when we optimize. After all, we usually cannot influence the true outcome $y$ or its distribution, only our point prediction $\hat{y}$.
Thus, I submit that it is more helpful to think about this as considering your loss as a random variable (through the uncertainty in the outcome $y$) which depends on a variable $\hat{y}$, and trying to understand which value of $\hat{y}$ minimizes, e.g., the expectation of your loss.
I like to investigate things like this through simple simulations. For example, in analogy to your other thread (Loss function that penalizes wrong sign predictions), assume your predictive uncertainty about the outcome can be parameterized as a normal distribution with mean 1 and standard deviation 2. It turns out that the $\hat{y}$ that minimizes the expected loss for $\alpha=1$ is 1.5, i.e., your loss incentivizes us to give a prediction that is higher than the mean. This may well be what you want - but it is important to be aware of this effect.

R code:
mm <- 1
sd <- 2
xx <- mm+seq(-2*sd,2*sd,by=0.01)
sims <- rnorm(1e6,mm,sd)
alpha <- 1
loss <- sapply(xx,function(yy)mean((sims-yy)^2-alphasimsyy))
xx[which.min(loss)]
plot(xx,loss,type="l",xlab="Point prediction",las=1,ylab="Expected loss")
abline(v=mm,col="red")