Parameter estimation is an essential part of statistical inference where we use our data to guess the value of an unknown variable. The true goal of parameter estimation is not simply to estimate a value, but rather to assign a probability for a range of possible values.
The normal distribution is a continuous distribution (like beta distribution) that describe the strength of possible beliefs in the value of an uncertain measurement, given a known mean and standard deviation. The PDF of the normal distribution is:
$$N(\mu, \sigma) = \frac{1}{\sqrt{2\pi \sigma^2}} \times e^{ -\frac{ (x - \mu)^2 }{ 2\sigma^2 } }$$
Mean of a beta distribution is given by:
$$\mu_{\text{Beta}} = \frac{\alpha}{\alpha + \beta}$$
The PDF is most useful visually for quickly estimating where the peak of a distribution is, and for getting a rough sense of the width (variance) and shape of a distribution. However, with PDF it is very difficult to reason about the probability of various ranges visually, the CDF is a much better tool for this.
Inverse of CDF is quantile function.