Probability mass at zero
How does the normal distribution have a zero probability mass at zero?
The normal distribution has a non zero density at zero but the probability (mass) is zero $P[X=0] = 0$.
By placing a probability mass at zero the prior is expressing more strongly the believe that a parameter is probably zero. That is helpful in a setting where many regressors are included, of which we believe most are not truly in the model.
Fatter tails
By having fatter tails we allow for a few of the parameters to have more easily larger values. In RIDGE and LASSO the penalty is not only keeping out the 'unwanted' overfitting of noise due to too many parameters, it also makes that the 'correct' model parameters are shrunken. The estimated parameter values with penalization are smaller than with the unbiased ordinary least squares estimate.
What's the difference and which is better?
So you can see this prior as a more extreme variant of LASSO in comparison to ridge, placing even more focus on parameter selection and less on regularising by shrinking parameters.
Note that the one is not neccesarily better than the other. Shrinkage is not always unwanted and regularisation is not all about parameter selection. They are just placing a different focus.
The horseshoe
The name “horseshoe” came from the shape of the distribution if we re-parametrize it using this transformation using shrinkage weight k:
What is this 'shrinkage weight'? It relates to the use of the following prior model:
$$\begin{array}{}
\beta_i &\sim& N(0,\tau \lambda_i) \\
\tau &\sim& \text{Half-Cauchy}(0,\tau_0) \\
\lambda_i &\sim& \text{Half-Cauchy}(0,1)
\end{array}$$
or in reparameterized form
$$\begin{array}{}
\beta_i &\sim& N\left(0,\frac{\tau}{\sqrt{\kappa_i^{-1}-1}} \right) \\
\tau &\sim& \text{Half-Cauchy}(0,\tau_0) \\
\kappa_i &\sim& \text{Beta}\left(\frac{1}{2},\frac{1}{2}\right)
\end{array}$$
The relationship between the beta distribution and the half-cauchy distribution (whose square is F distributed), can be seen when we rewrite the reparameterization as $\lambda^2 = \frac{1-\kappa}{\kappa}$, and that resembles the transformation between the F-distribution and the beta distribution written in several places (e.g. Wikipedia here).
This $\kappa_i$ relates to the size of the prior $N\left(0,\frac{\tau}{{\kappa_i^{-1}-1}} \right)$ and makes it either a point mass when $\kappa = 1$ or a heavy tailed distribution when $\kappa = 0$.