This is a great question because it is a variation of a problem/issue my students often pose when learning about confidence and prediction interval estimation with multiple regression (MR). The addition of the bootstrapping element for the parameters to the simulation protocol is indeed appropriate (and I'm surprised I haven't thought about it for class/lecture demonstration purposes).
My hope is to provide a useful explanation for why we definitely should bootstrap the parameters in order to simulate the response distribution. But, I will also provide a simpler protocol (which draws on the prediction interval from an MR analysis). And lastly, I'll briefly indicate a scenario where the bootstrapping protocol detailed here would probably be more applicable.
First, in the OP, you find the statement: “The true response $Y^+ = \beta x_+ + \epsilon$ does not vary for different realisations of our sample data, contrary to the predictor $Y^+$, and so the variation in the parameter estimate should not be relevant to its simulated distribution.” This actually is not a fully correct statement...but part of it is...and this is the key to understanding the need for the simulation.
Let me use the convention that lower-case Greek letters represent the parameter (true) values of the population, and upper-case Latin letters represent sample estimates for these parameters. So, while it is true for all values of the population that
$$Y =\beta x + \epsilon$$
it is key to remember that when we obtain our parameter estimates from a sample, we do not have $\beta$, but $B$. And even if these values are off by just a little bit, this means the error estimate will also be incorrect. So, what we have is
$$Y = B x + E$$
If you think about this as a basic bivariate regression, if your slope estimate is off from the population slope by just a little bit, well...you can still find the error to make your predicted values match your observed values of $Y$, but one of either the predicted value or the error estimate will be too high (and the other too low). So, if you just run a bootstrap using the estimated slope, you run the risk of making all of your simulated predictions be a little bit too high or low.
Thus, in a bootstrapping approach, we indeed would want to simulate a variety of the possible $B$ estimates we might obtain in order to make our prediction distribution.
And the nice part is that the normal theory behind ordinary least squares (OLS) MR has already answered this question of future prediction without needing to rely on bootstrapping the parameter estimates. Without elaborating the proof of these intervals here, I will give the confidence interval for the conditional mean and the prediction interval.
The confidence interval for the conditional mean is a range of values in which we would reasonably expect to find the mean value of $Y$ for some given value $x$ (that is to say, $\mu_Y$ is conditioned on knowing this value of $x$). This interval is
$$\mu_{Y|x_+}= \hat{Y}_+ \pm t_\text{c.v.} · \hat{\sigma}_\epsilon\sqrt{x_+ (X^T X)^{-1} x_+^T}$$
This interval adds variability to our predicted (estimated) value to account for the fact that our parameter estimates for the regression $B$ may not have perfectly matched up with the population parameters $\beta$ in our MR model.
Next, we can adjust this formula slightly to get the prediction interval for any future value of $Y$ at this value of $x_+$:
$$Y|x_+= \hat{Y}_+ \pm t_\text{c.v.} · \hat{\sigma}_\epsilon\sqrt{1+ x_+ (X^T X)^{-1} x_+^T}$$
This interval accounts for the variability in the parameter estimates from the MR analysis, and it accounts for the error variability (which also has been estimated from the MR analysis).
And this gives us the simpler bootstrapping protocol. All we need do is simulate new responses at the point of interest $x_+$ by drawing $Y_+^{(b)}$ from the normal distribution with mean $\hat{Y}_+$ and standard deviation $\hat{\sigma}_\epsilon\sqrt{1+ x_+ (X^T X)^{-1} x_+^T}$.
However, as noted above...as this is a fairly well understood problem (and solution), I am unsure why the bootstrapping protocol would be necessary unless there is some violation of the assumptions for the MR model. And this brings me to my final point: ¿when would this bootstrapping protocol be appropriate?
I would argue that this protocol would be appropriate if you are using some other estimation method to obtain the regression model, like a robust estimation using the median instead of the mean. As the associated distributions of these protocols are more complicated, a bootstrapping process that resamples the parameter estimates prior to making each prediction would be appropriate to gain a better understanding of the distribution of future predicted values.
I hope this answer proves useful, and I’m happy to elaborate further as needed.