Recall that the likelihood of a model is the probability of the data set given the model (P(D|θ)).
The deviance of a model is defined by
D(θ, D) = 2(log (P(D|θs)) − log (P(D|θ)))
where θs is the saturated model which is so named because it perfectly fits the data.
In the case of normally distributed errors the likelihood for a single prediction (μi) and data point (yi) is given by
$$P(y_i|\mu_i) = \frac{1}{\sigma\sqrt{2\pi}} \exp\bigg(-\frac{1}{2}\bigg(\frac{y_i - \mu_i}{\sigma}\bigg)^2\bigg)$$ and the log-likelihood by
$$\log(P(y_i|\mu_i)) = -\log(\sigma) - \frac{1}{2}\big(\log(2\pi)\big) -\frac{1}{2}\bigg(\frac{y_i - \mu_i}{\sigma}\bigg)^2$$
The log-likelihood for the saturated model, which is when μi = yi, is therefore simply
$$\log(P(y_i|\mu_{s_i})) = -\log(\sigma) - \frac{1}{2}\big(\log(2\pi)\big)$$
It follows that the unit deviance is
di = 2(log (P(yi|μsi)) − log (P(yi|μi)))
$$d_i = 2\bigg(\frac{1}{2}\bigg(\frac{y_i - \mu_i}{\sigma}\bigg)^2\bigg)$$
$$d_i = \bigg(\frac{y_i - \mu_i}{\sigma}\bigg)^2$$
As the deviance residual is the signed squared root of the unit deviance,
$$r_i = \text{sign}(y_i - \mu_i) \sqrt{d_i}$$ in the case of normally distributed errors we arrive at $$r_i = \frac{y_i - \mu_i}{\sigma} $$ which is the Pearson residual.
To confirm this consider a normal distribution with a μ̂ = 2 and σ = 0.5 and a value of 1.