While we often use F-testing for testing differences in means (e.g., ANOVA), the F-test is actually a test of variances that methods like ANOVA use cleverly to investigate differences in means.
Therefore, the first thought might be to use an F-test of the two variances. This can be implemented in R software, for instance, using var.test.
Unfortunately, the F-test lacks robustness to deviations from normality. The JBStatistics channel on YouTube has a video showing this, and it might be fun to come up with your own simulations to show this.
A more robust alternative is the Ansari-Bradley test, implemented in R through ansari.test. Technically, this is not quite a variance test, but it tends to do a good job and could be worth a read.
If you want to get into a more general setting where you find the variance conditional on multiple covariates, this question of mine is asking the same and has yet to get the kind of resolution I had hoped to get.
For quantifying the effect size, I find it natural to talk about the ratio of the two variances, rather than the difference. It makes sense to me to say that one distribution has twice or half the variance of another, and this ratio is part of what is calculated in the F-test.
Finally, establishing causality is likely to encounter the same kind of bugaboos that occur when it comes to establishing causality in a regression that estimates conditional means. This is good, because people who do causal inference already have tools to do so (e.g., instrumental variables), yet the estimation is different (estimating a conditional variance instead of a conditional mean), so the theoretical motivation in the causal inference may be more difficult, and the techniques may not be as well established with easy availability in software (e.g., the analogue to instrumental variables when conditional variances are being estimated).