I have recently been studying Bayesian as well as Frequentive Statistics (mostly null hypothesis significance testing) and am confused as to the meaning of the distribution of the likelihood and observed data in both. According to my understanding, Bayesian school of though treats the distribution of the sampled data to be same as that of the likelihood of th data conditioned over the hypothesis (For example if data follows a normal distribution then the likelihood function is also normal with same mean and variance) whereas in the frequentive school, the likelihood or null distribution is basically the distribution followed by the test statistic which is different from the distribution followed by the data. Could someone explain whether my understanding is correct?
2026-03-31 19:13:34.1774984414
Probability distribution of data and likelihood in Bayesian and Frequentive Statistics
238 Views Asked by Bumbble Comm https://math.techqa.club/user/bumbble-comm/detail At
1
There are 1 best solutions below
Related Questions in PROBABILITY
- How to prove $\lim_{n \rightarrow\infty} e^{-n}\sum_{k=0}^{n}\frac{n^k}{k!} = \frac{1}{2}$?
- Is this a commonly known paradox?
- What's $P(A_1\cap A_2\cap A_3\cap A_4) $?
- Prove or disprove the following inequality
- Another application of the Central Limit Theorem
- Given is $2$ dimensional random variable $(X,Y)$ with table. Determine the correlation between $X$ and $Y$
- A random point $(a,b)$ is uniformly distributed in a unit square $K=[(u,v):0<u<1,0<v<1]$
- proving Kochen-Stone lemma...
- Solution Check. (Probability)
- Interpreting stationary distribution $P_{\infty}(X,V)$ of a random process
Related Questions in PROBABILITY-THEORY
- Is this a commonly known paradox?
- What's $P(A_1\cap A_2\cap A_3\cap A_4) $?
- Another application of the Central Limit Theorem
- proving Kochen-Stone lemma...
- Is there a contradiction in coin toss of expected / actual results?
- Sample each point with flipping coin, what is the average?
- Random variables coincide
- Reference request for a lemma on the expected value of Hermitian polynomials of Gaussian random variables.
- Determine the marginal distributions of $(T_1, T_2)$
- Convergence in distribution of a discretized random variable and generated sigma-algebras
Related Questions in STATISTICS
- Given is $2$ dimensional random variable $(X,Y)$ with table. Determine the correlation between $X$ and $Y$
- Statistics based on empirical distribution
- Given $U,V \sim R(0,1)$. Determine covariance between $X = UV$ and $V$
- Fisher information of sufficient statistic
- Solving Equation with Euler's Number
- derive the expectation of exponential function $e^{-\left\Vert \mathbf{x} - V\mathbf{x}+\mathbf{a}\right\Vert^2}$ or its upper bound
- Determine the marginal distributions of $(T_1, T_2)$
- KL divergence between two multivariate Bernoulli distribution
- Given random variables $(T_1,T_2)$. Show that $T_1$ and $T_2$ are independent and exponentially distributed if..
- Probability of tossing marbles,covariance
Related Questions in STATISTICAL-INFERENCE
- co-variance matrix of discrete multivariate random variable
- Question on completeness of sufficient statistic.
- Probability of tossing marbles,covariance
- Estimate the square root of the success probability of a Binomial Distribution.
- A consistent estimator for theta is?
- Using averages to measure the dispersion of data
- Confidence when inferring p in a binomial distribution
- A problem on Maximum likelihood estimator of $\theta$
- Derive unbiased estimator for $\theta$ when $X_i\sim f(x\mid\theta)=\frac{2x}{\theta^2}\mathbb{1}_{(0,\theta)}(x)$
- Show that $\max(X_1,\ldots,X_n)$ is a sufficient statistic.
Related Questions in BAYESIAN
- Obtain the conditional distributions from the full posterior distribution
- What it the posterior distribution $\mu| \sigma^2,x $
- Posterior: normal likelihood, uniform prior?
- If there are two siblings and you meet one of them and he is male, what is the probability that the other sibling is also male?
- Aggregating information and bayesian information
- Bayesian updating - likelihood
- Is my derivation for the maximum likelihood estimation for naive bayes correct?
- I don't understand where does the $\frac{k-1}{k}$ factor come from, in the probability mass function derived by Bayesian approach.
- How to interpret this bayesian inference formula
- How to prove inadmissibility of a decision rule?
Trending Questions
- Induction on the number of equations
- How to convince a math teacher of this simple and obvious fact?
- Find $E[XY|Y+Z=1 ]$
- Refuting the Anti-Cantor Cranks
- What are imaginary numbers?
- Determine the adjoint of $\tilde Q(x)$ for $\tilde Q(x)u:=(Qu)(x)$ where $Q:U→L^2(Ω,ℝ^d$ is a Hilbert-Schmidt operator and $U$ is a Hilbert space
- Why does this innovative method of subtraction from a third grader always work?
- How do we know that the number $1$ is not equal to the number $-1$?
- What are the Implications of having VΩ as a model for a theory?
- Defining a Galois Field based on primitive element versus polynomial?
- Can't find the relationship between two columns of numbers. Please Help
- Is computer science a branch of mathematics?
- Is there a bijection of $\mathbb{R}^n$ with itself such that the forward map is connected but the inverse is not?
- Identification of a quadrilateral as a trapezoid, rectangle, or square
- Generator of inertia group in function field extension
Popular # Hahtags
second-order-logic
numerical-methods
puzzle
logic
probability
number-theory
winding-number
real-analysis
integration
calculus
complex-analysis
sequences-and-series
proof-writing
set-theory
functions
homotopy-theory
elementary-number-theory
ordinary-differential-equations
circles
derivatives
game-theory
definite-integrals
elementary-set-theory
limits
multivariable-calculus
geometry
algebraic-number-theory
proof-verification
partial-derivative
algebra-precalculus
Popular Questions
- What is the integral of 1/x?
- How many squares actually ARE in this picture? Is this a trick question with no right answer?
- Is a matrix multiplied with its transpose something special?
- What is the difference between independent and mutually exclusive events?
- Visually stunning math concepts which are easy to explain
- taylor series of $\ln(1+x)$?
- How to tell if a set of vectors spans a space?
- Calculus question taking derivative to find horizontal tangent line
- How to determine if a function is one-to-one?
- Determine if vectors are linearly independent
- What does it mean to have a determinant equal to zero?
- Is this Batman equation for real?
- How to find perpendicular vector to another vector?
- How to find mean and median from histogram
- How many sides does a circle have?
Your terminology is not exactly on the right track. Here is an example to show the likelihood function and how it is used in frequentist and Bayesian estimation.
Bayesian and Frequentist Inference for Bernoulli Data
Suppose we have a random sample $X_1, \dots, X_n$ from a certain population. In both frequentist and Bayesian approaches, the likelihood has the same form as the JOINT density function of $X_1, \dots, X_n$. Not the density function of the population (or of a single observation).
$Likelihood.$ In its role as a likelihood function, the joint density is viewed as a function of (one or more) unknown parameters, with observed data values $x_1, \dots, x_n$ (often now written in lower case) regarded as fixed known values. Another convention for the likelihood function is to say that it is defined only up to a constant multiple.
For example, suppose the $X_i$ are Bernoulli with parameter $\theta$ (success probability). Then the joint 'density' function is
$$f(\text{x}|\theta) = \prod_{i=1}^n \theta^{X_i}(1 - \theta)^{1 - X_i} = \theta^S(1-\theta)^{n-S},$$ where $S = \sum_{i=1}^n X_i.$ We often write $f(\text{x}|\theta) \propto \theta^S(1-\theta)^{n-S},$ where the proportionality symbol $\propto$ indicates that we may leave out factors that do not involve the parameter of interest.
Then $S \sim Binom(n, \theta).$ Specifically, if $n = 1000,$ and we observe $x_i$ with total $s = 620,$ then $f(s|\theta) = {1000 \choose 620}\theta^{620}(1-\theta)^{380} \propto \theta^{620}(1-\theta)^{380}.$
$Frequentist.$ From a frequentist point of view, we find the maximum likelihood estimator $\hat \theta = s/n = 620/1000 = 0.62.$ (The point at which $f(s|\theta$ reaches its maximum is $s/n$.)
Various confidence intervals have been proposed. A common 95% CI (based on normal and other approximations and not so good for small $n$) is $\hat \theta \pm 1.96\sqrt{\hat \theta(1-\hat \theta)/n}.$ For the data of our example, this is $(0.590, 0.650).$
The usual frequentist interpretations are that $\hat\theta$ has good properties as an estimator of $\theta$: it is unbiased because $E(\hat \theta) = \theta$ and it has minimum variance among unbiased estimators. The CI is said to have been produced by a procedure by which in 95% of such experiments over the long run, the CI will cover the fixed, but unknown, constant $\theta.$
$Bayesian.$ A Bayesian analysis view $\theta$ as a random variable with a prior distribution. Because $\theta$ must lie in the unit interval, a member of the Beta family of distributions is a common choice for the prior distribution of $\theta:$ perhaps $Beta(1/2, 1/2)$ or $Beta(1,1)$ if we have no previous experience or strong personal opinion about $\theta;$ perhaps something like $Beta(330, 270)$ if we think $\theta$ is likely to lie between 0.51 and 0.59, based on previous experiments, polls or hunches about the population.
Then Bayes Theorem comes into play. It states that the posterior distribution of $\theta$ based on the prior distribution and the likelihood function is $$f(\theta|s) \propto f(\theta) \times f(s|\theta),$$
where the (last suggested) prior distribution has 'kernel' $f(\theta) \propto \theta^{330-1}(1 - \theta^{270-1})$ and the likelihood (from above) is $f(s|\theta) \propto \theta^{620}(1-\theta)^{380}.$
By the 'kernel' of a distribution we mean the factors that involve the parameter of concern. Thus the kernel of the posterior distribution is $f(\theta|x) \propto \theta^{950-1}(1-\theta)^{650-1}$ and the posterior distribution is obviously $Beta(950, 650).$ One possible 95% Bayesian probability interval for $\theta$ cuts 2.5% from each tail of this distribution to give $(0.570, 0.618),$ which can be obtained from R code
qbeta(c(.025,.975), 950, 650). A Bayesian point estimate might be the mean of the posterior distribution $950/(950+650) = 0.594$ (or its mode, which is essentially the same in this case).The Bayesian interpretation of these results is that--for the particular survey or experiment at hand--the most likely value of $\theta$ is 59.4% and that there is 95% probability the true value of $\theta$ is between 57.0% and 61.8%. (Of course these statements are subject to believing the prior distribution.)
If our prior distribution had been $Beta(1,1) = Unif(0,1),$ (a 'non-informative' prior) the 95% probability interval would have been $(0.589, 0.650)$, numerically similar to the frequentist interval above.
Notes: (1) For simplicity, I have used 'density' for both discrete and continuous distributions; for former, I might have used 'point mass function'. (2) Bayesian statisticians often use $p$ instead of $f$ for density and likelihood functions. functions. (3) Finding the posterior distribution was very simple in the example given, because of the algebraic compatibility of the prior and likelihood. In such cases we say the prior and likelihood are 'conjugate'. Finding probability intervals for some posterior distributions requires intensive computation.