Eu acho que é mais correto falar da distribuição posterior do seu parâmetro vez de sua estimativa posterior. Para maior clareza das anotações, deixarei o primo em seguir.σ′2σ′2
Suponha que seja distribuído como , - eu largo por enquanto para fazer um exemplo heurístico - e é distribuído como e é independente de .XN(0,σ2)μ1/σ2=σ−2Γ(α,β)X
O pdf de dado é gaussiano, ieXσ−2
f(x|σ−2)=12πσ2−−−−√exp(−x22σ2).
O pdf conjunto de , é obtido pela multiplicação de por - o pdf de . Isso sai como(X,σ−2)f(x,σ−2)f(x|σ−2)g(σ−2)σ−2
f(x,σ−2)=12πσ2−−−−√exp(−x22σ2)βαΓ(α)exp(−βσ2)1σ2(α−1).
We can group similar terms and rewrite this as follows
f(x,σ−2)∝σ−2(α−1/2)exp(−σ−2(β+x2/2)).
The posterior distribution of σ−2 is by definition the pdf of σ−2 given x, which is f(x,σ−2)/f(x) by Bayes' formula. To answer your question 1. I don't think there is a way to express f(σ−2|x) from f(x,σ−2) without using Bayes' formula. On with the computation, we recognize in the formula above something that looks like a Γ function, so integrating σ−2 out to get f(x) is fairly easy.
f(x)∝(β+x2/2)−(α+1/2),
so by dividing we get
f(σ−2|x)∝(β+x2/2)(σ−2(β+x2/2))α−1/2exp(−σ−2(β+x2/2))∝(σ−2(β+x2/2))α−1/2exp(−σ−2(β+x2/2)).
And here in the last formula we recognize a Γ distribution with parameters (α+1/2,β+x2/2).
If you have an IID sample ((x1,σ−21),...,(xn,σ−2n)), by integrating out all the σ−2i, you would get f(x1,...,xn) and then f(σ−21,...,σ−2n|x1,...,xn) as a product of the following terms:
f(σ−21,...,σ−2n|x1,...,xn)∝∏i=1n(σ−2i(β+x2i/2))α−1/2exp(−σ−2i(β+x2i/2)),
Which is a product of Γ variables. And we are stuck here because of the multiplicity of the σ−2i. Besides, the distribution of the mean of those independent Γ variables is not straightforward to compute.
However, if we assume that all the observations xi share the same value of σ−2 (which seems to be your case) i.e. that the value of σ−2 was drawn only once from a Γ(α,β) and that all xi were then drawn with that value of σ−2, we obtain
f(x1,...,xn,σ−2)∝σ−2(α+n/2)exp(−σ−2(β+12∑i=1nx2i)),
from which we derive the posterior distribution of σ−2 as your equation 1 by applying Bayes' formula.
The posterior distribution of σ−2 is a Γ that depends on α and β, your prior parameters, the sample size n and the observed sum of squares. The prior mean of σ−2 is α/β and the variance is α/β2, so if α=β and the value is very small, the prior carries very little information about σ−2 because the variance becomes huge. The values being small, you can drop them from the above equations and you end up with your equation 3.
In that case the posterior distribution becomes independent of the prior. This formula says that the inverse of the variance has a Γ distribution that depends only on the sample size and the sum of squares. You can show that for Gaussian variables of known mean, S2, the estimator of the variance, has the same distribution, except that it is a function of the sample size and the true value of the parter σ2. In the Bayesian case, this is the ditribution of the parameter, in the frequentist case, it is the distribution of the estimator.
Regarding your question 2. you can of course use the values obtained in a previous experiment as your priors. Because we established a parallel between Bayesian and frequentist interpretation in the above, we can elaborate and say that it is like computing a variance from a small sample size and then collecting more data points: you would update your estimate of the variance rather than throw away the first data points.
Regarding your question 3. I like the Introduction to Mathematical Statistics by Hogg, McKean and Craig, which usually gives the detail of how to derive these equations.