Questions tagged [bayesian]

Bayesian inference is a method of statistical inference that relies on treating the model parameters as random variables and applying Bayes' theorem to deduce subjective probability statements about the parameters or hypotheses, conditional on the observed dataset.

Overview

Bayesian inference is a method of statistical inference that treats model parameters as if they were random variables in order to rely on probability calculus and produces complete and unified probabilistic statements about these parameters.

This approach starts with choosing a reference or prior probability distribution on the parameters and then applies Bayes' Theorem to deduce probability statements about parameters or hypotheses, conditional on the data, treating the likelihood function as a conditional density of the data given the (random) parameter. Bayes' Theorem asserts that the conditional density of the parameter $\theta$ given the data, $P(\theta|d)$, can be expressed in terms of the density of the data given $\theta$ as

$$P(\theta|d) = \dfrac{P(d|\theta)P(\theta)}{P(d)}.$$

$P(\theta|d)$ is called the posterior probability. $P(d|\theta)$ is often called the likelihood function and denoted $L(\theta|d)$. The distribution of $\theta$ itself, given by $P(\theta)$, is called the prior or the reference measure. It encodes previous or prior beliefs about $\theta$ within a model appropriate for the data. There is necessarily a part of arbitrariness or subjectivity in the choice of that prior, which means that the resulting inference is impacted by this choice (or conditional to it). This also means that two different choices of priors lead to two different posterior distributions, which are not directly comparable.

The marginal distribution of the data, $P(d)$ (which appears as a normalization factor), is also called the evidence, as it is directly used for Bayesian model comparison through the notions of Bayes factors and model posterior probabilities.

The comparison of two models (including two opposed hypotheses about the parameters) in the Bayesian framework indeed proceeds by taking the ratio of the evidences for these two models under comparisons, $$ B_{12} = P_1(d)\big/P_2(d)\,. $$ This is called the Bayes factor and it is usually compared to $1$.

Bayes' formula can be used as an updating procedure: as more data become available, the posterior can be updated successively, becoming the prior for the next step.

References

The following threads contain lists of references:

The following journal is dedicated to research in Bayesian statistics:

8003 questions
50
votes
8 answers

What are the cons of Bayesian analysis?

What are some practical objections to the use of Bayesian statistical methods in any context? No, I don't mean the usual carping about choice of prior. I'll be delighted if this gets no answers.
user6666
28
votes
3 answers

What is the Bayesian justification for privileging analyses conducted earlier than other analyses?

Background and Empirical Example I have two studies; I ran an experiment (Study 1) and then replicated it (Study 2). In Study 1, I found an interaction between two variables; in Study 2, this interaction was in the same direction but not…
Mark White
  • 10,252
27
votes
3 answers

Why break down the denominator in Bayes' Theorem?

(I'm a newbie at stats. I'm a mathematician and a programmer and I'm trying to build something like a naive Bayesian spam filter.) I've noticed in many places that people tend to break down the denominator in the equation from Bayes' Theorem. So…
24
votes
6 answers

"Fully Bayesian" vs "Bayesian"

I have been learning about Bayesian statistics, and I often have read in articles "we adopt a Bayesian approach" or something similar. I also noticed, less often: "we adopt a fully Bayesian approach" (my emphasis). Is there any difference…
Joe King
  • 3,805
13
votes
2 answers

Conjugate prior for a Gamma distribution

I need to update the failure rate (given as deterministic) based on new rate of failure about the same system (it is a deterministic one too). I read about conjugate priors and Gamma distribution as a conjugate for the Poisson process. Also, I can…
Alone
  • 141
12
votes
1 answer

Bayesian updating - coin tossing example

I have a question about Bayesian updating. In general Bayesian updating refers to the process of getting the posterior from a prior belief distribution. Alternatively one could understand the term as using the posterior of the first step as prior…
TinaW
  • 393
11
votes
4 answers

Is the "Bayesian approach" about prior beliefs, viewing parameters as random variables, or both?

It seems to me like the concepts of incorporating prior beliefs about parameters VERSUS viewing parameters as latent random variables are two VERY separate concepts, and yet I've found that they're often confounded and treated as one. Can't we have…
11
votes
1 answer

Examples of misapplication of Bayes theorem

This Math Overflow community question asked for "examples of bad arguments which involve the application of mathematical theorems in non-mathematical contexts" and produced a fascinating list of pathologically applied mathematics. I am wondering…
10
votes
3 answers

Posterior Predictive Distribution - Normal model with unknown mean, and known std dev

I do not understand how one can prove that in a normal model, where we know the variance, but not the mean, the posterior predictive distribution is also normal... Below is cut of a page from Gelman's Data Bayesian Analysis. I don't understand how…
9
votes
1 answer

Non-conjugate prior

Can someone explain why the integral in the posterior density may not be "analytically tractable" if the prior we choose is non-conjugate?
user35401
9
votes
2 answers

How does the marginal distribution become the prior distribution?

Bayes' Theorem for densities/pmf's states that, given, say, two univariate random variables $x,z$ we have $$p(z\mid x) = \frac {p(x\mid z)\cdot p(z)}{p(x)}$$ This is part of the core of the probability theory most widely used nowadays. But in order…
9
votes
1 answer

Can a proper prior lead to an improper posterior?

It is my understanding that if we have an improper prior, we may either get a proper or a posterior distribution. My question is if there are cases where a proper prior distribution may, when combined with the likelihood function, result in an…
user1398057
  • 2,365
  • 1
  • 20
  • 24
9
votes
2 answers

Arithmetic for updating likelihoods using Bayes theorem

This may be an elementary question which is why I have not been able to find it on Stackexchange or Mathoverflow however I am having problems with the arithmetic involved in updating likelihoods using Bayes theorem for a problem I am working…
Andrew
  • 93
9
votes
1 answer

Bayesian rating system with multiple categories for each rating

I'm implementing a rating system to be used on my website, and I think the Bayesian average is the best way to go about it. Every item will be rated in six different categories by the users. I don't want items with only one high rating to shoot to…
8
votes
3 answers

Counter intuitive Bayesian theorem

Consider an example as follows. I am running a mobile app that allows users to apply for a loan on the app. Say a guy signed in to my app to use his phone to apply for a loan. Call events: A = a person has a smartphone B = default First of all, I…
mommomonthewind
  • 987
  • 2
  • 11
  • 21
1
2 3
14 15