As whuber points out in the comments, there are a range of possible answers for how we might estimate in this case, but ultimately we would then need to investigate the quality of the proposed estimators by looking at their statistical properties (e.g., bias, consistency, MSE, etc.). Consequently, the only real way to answer this question is to set out some proposed estimators for this case and suggest that their properties be examined. The general problem you will encounter is that an "adjustment" used to incorporate the information for the order of the distribution means is likely to bias the estimators, though it might lead to some other beneficial properties. Ultimately, you would need to investigate the properties of any proposed estimators and compare them to standard estimators to decide if they are an improvement.
If we want to proceed in the greatest generality, without assuming any particular distributional form, then a natural starting point is to note that for IID data we usually estimate the distribution mean with the sample mean. Consequently, if we have IID samples from the two distributions (and they are assumed to also be independent of each other) then a natural starting point would be to estimate the distribution means with the sample means. However, if the sample means are in the "wrong" order compared to the known order of the distribution means, we might then propose to adjust our estimator by estimating both distribution means as some point between the two sample means.
Let's proceed on this basis and make a small start on formulating proposed estimators and examining a basic property of those estimators. Suppose we have $n$ observations $X_1,...,X_n$ and $m$ observations $Y_1,...,Y_m$. Using some proportion $0 \leqslant p(n,m) \leqslant 1$ we might decide to form the estimators:
$$\hat{\mu}_X = \begin{cases}
\bar{X}_n & & & \text{if } \bar{X}_n < \bar{Y}_m, \\[6pt]
p(n,m) \bar{X}_n + (1-p(n,m)) \bar{Y}_m & & & \text{if } \bar{X}_n \geqslant \bar{Y}_m. \\[6pt]
\end{cases} \\[24pt]
\hat{\mu}_Y = \begin{cases}
\bar{Y}_m & & & \text{if } \bar{X}_n < \bar{Y}_m, \\[6pt]
p(n,m) \bar{X}_n + (1-p(n,m)) \bar{Y}_m & & & \text{if } \bar{X}_n \geqslant \bar{Y}_m. \\[6pt]
\end{cases}$$
It can easily be shown that $\hat{\mu}_X \leqslant \hat{\mu}_Y$ so that the estimators respect the known order of the distribution means. To find the expected value of the estimators, suppose we define the quantities:
$$\begin{align}
K(n,m) &\equiv \mathbb{P}(\bar{X}_n < \bar{Y}_m), \\[6pt]
L_X(n,m) &\equiv \mathbb{E}(\bar{X}_n|\bar{X}_n < \bar{Y}_m), \\[6pt]
L_Y(n,m) &\equiv \mathbb{E}(\bar{Y}_m|\bar{X}_n < \bar{Y}_m), \\[6pt]
U_X(n,m) &\equiv \mathbb{E}(\bar{X}_n|\bar{X}_n \geqslant \bar{Y}_m), \\[6pt]
U_Y(n,m) &\equiv \mathbb{E}(\bar{Y}_m|\bar{X}_n \geqslant \bar{Y}_m), \\[6pt]
\end{align}$$
and note the operative constraints:
$$\begin{align}
\mu_X = L_X(n,m) \cdot K(n,m) + U_X(n,m) \cdot (1-K(n,m)), \\[6pt]
\mu_Y = L_Y(n,m) \cdot K(n,m) + U_Y(n,m) \cdot (1-K(n,m)). \\[6pt]
\end{align}$$
Using these quantities we have the expected values:
$$\begin{align}
\mathbb{E}(\hat{\mu}_X)
&= \mathbb{E}(\hat{\mu}_X|\bar{X}_n < \bar{Y}_m) K(n,m) + \mathbb{E}(\hat{\mu}_X|\bar{X}_n \geqslant \bar{Y}_m) (1-K(n,m)) \\[6pt]
&= L_X(n,m) K(n,m) + U_X(n,m) p(n,m) (1-K(n,m)) \\[6pt]
&\quad \quad + U_Y(n,m) (1-p(n,m)) (1-K(n,m)), \\[12pt]
\mathbb{E}(\hat{\mu}_Y)
&= \mathbb{E}(\hat{\mu}_Y|\bar{X}_n < \bar{Y}_m) K(n,m) + \mathbb{E}(\hat{\mu}_Y|\bar{X}_n \geqslant \bar{Y}_m) (1-K(n,m)) \\[6pt]
&= L_Y(n,m) K(n,m) + U_X(n,m) p(n,m) (1-K(n,m)) \\[6pt]
&\quad \quad + U_Y(n,m) (1-p(n,m)) (1-K(n,m)). \\[6pt]
\end{align}$$
We therefore have:
$$\begin{align}
\text{Bias}(\hat{\mu}_X)
&= [U_Y(n,m) - U_X(n,m)] (1-p(n,m)) (1-K(n,m)), \\[12pt]
\text{Bias}(\hat{\mu}_Y)
&= [U_X(n,m) - U_Y(n,m)] p(n,m) (1-K(n,m)). \\[6pt]
\end{align}$$
As you can see, these estimators would generally be biased, though they have the advantage of respecting the known ordering in the underlying distribution means. The amount of bias will depend on probabilities and conditional moments from the underlying distributions (through the quantities $U_X$, $U_Y$ and $K$). The bias also depends on the proportion $p(n,m)$ used in the estimator, which is under our control. In particular, if we take $p(n,m) = \tfrac{1}{2}$ then the magnitude of the bias will be the same for both estimators.
With some more analysis you could investigate other properties of this estimator, or formulate alternative estimators and then examine their properties. Here we have just scratched the surface of this type of analysis, but hopefully it gives you an idea of how you might proceed with this type of analysis.