Can anyone give a concrete example to illustrate what is an uniform prior?Given an observed sample from a binomial distribution and a known prior, how can I update the probability distribution of possible 'p' values?Probablity of getting sequence of K equal results while tossing coin N timesWhy doesn't ML point estimate equal MAP point estimate even though I'm using uniform prior?Long-run behavior in coin tossing experimentProbability of $k$ successes in no more than $n$ Bernoulli trialsParadoxical results when performing simple Bayesian analysis. Need help explain resultsHow to calculate probability distribution of rolling n dice? (with a twist!)How does a Bayesian update his belief when something with probability 0 happened?Bayesian posterior pmf for weighted dice with uniform priorWhat is the likelihood function of having heads 8 times out of 10 tossHypothesis test between two coins
What are the different qualities of the intervals?
How to tar a list of directories only if they exist
What do you call a flexible diving platform?
Japanese reading of an integer
How do I explain an exponentially complex intuitively?
What do I do with a party that is much stronger than their level?
How did the Axis intend to hold the Caucasus?
Correlation length anisotropy in the 2D Ising model
Is it legal for private citizens to "impound" e-scooters?
Vertical tennis ball into fancy new enumerate
Are the named pipe created by `mknod` and the FIFO created by `mkfifo` equivalent?
Why is it considered Acid Rain with pH <5.6
How to kill my goat in Goat Simulator
Send a single HTML email from Thunderbird, overriding the default "plain text" setting
Why do all my history books divide Chinese history after the Han dynasty?
If a 2019 UA artificer has the Repeating Shot infusion on two hand crossbows, can they use two-weapon fighting?
How could Nomadic scholars effectively memorize libraries worth of information
Did the IBM PC use the 8088's NMI line?
Why can't my huge trees be chopped down?
Assuring luggage isn't lost with short layover
What is the use of で in this sentence?
(2 of 11: Moon-or-Sun) What is Pyramid Cult's Favorite Camera?
Writing a clean implementation of rock–paper–scissors game in C++
Why does Canada require mandatory bilingualism in all government posts?
Can anyone give a concrete example to illustrate what is an uniform prior?
Given an observed sample from a binomial distribution and a known prior, how can I update the probability distribution of possible 'p' values?Probablity of getting sequence of K equal results while tossing coin N timesWhy doesn't ML point estimate equal MAP point estimate even though I'm using uniform prior?Long-run behavior in coin tossing experimentProbability of $k$ successes in no more than $n$ Bernoulli trialsParadoxical results when performing simple Bayesian analysis. Need help explain resultsHow to calculate probability distribution of rolling n dice? (with a twist!)How does a Bayesian update his belief when something with probability 0 happened?Bayesian posterior pmf for weighted dice with uniform priorWhat is the likelihood function of having heads 8 times out of 10 tossHypothesis test between two coins
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;
$begingroup$
I googled "uniform prior" and got a link to Prior probability, which uses the term without an explanation or a definition.
anther link is to quora, which does not give an concrete example.
Can anyone give a concrete example, such as coin flipping or dice tossing, to illustrate the uniform prior
?
probability bayesian mathematical-statistics
$endgroup$
add a comment |
$begingroup$
I googled "uniform prior" and got a link to Prior probability, which uses the term without an explanation or a definition.
anther link is to quora, which does not give an concrete example.
Can anyone give a concrete example, such as coin flipping or dice tossing, to illustrate the uniform prior
?
probability bayesian mathematical-statistics
$endgroup$
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago
add a comment |
$begingroup$
I googled "uniform prior" and got a link to Prior probability, which uses the term without an explanation or a definition.
anther link is to quora, which does not give an concrete example.
Can anyone give a concrete example, such as coin flipping or dice tossing, to illustrate the uniform prior
?
probability bayesian mathematical-statistics
$endgroup$
I googled "uniform prior" and got a link to Prior probability, which uses the term without an explanation or a definition.
anther link is to quora, which does not give an concrete example.
Can anyone give a concrete example, such as coin flipping or dice tossing, to illustrate the uniform prior
?
probability bayesian mathematical-statistics
probability bayesian mathematical-statistics
asked 9 hours ago
czlswsczlsws
1696 bronze badges
1696 bronze badges
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago
add a comment |
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
Let's say you don't know the probability of head, $p$, of a coin. You decide to conduct an experiment to estimate what it is, via Bayesian analysis. It requires you to choose a prior, and in general you're free to choose one of the feasible ones. If you don't know or don't want to assume anything about this $p$, you can say that it is uniformly distributed in $[0,1]$, in which $f_P(p)=1, 0leq pleq 1$, $0$ otherwise. This is quite similar to saying that any $p$ value in $[0,1]$ is equally likely. This prior distribution is a uniform prior.
You can also choose other priors, that focus on different regions in $[0,1]$, for example, if you choose a prior like $f_P(p)=frac32(1-(1-2p)^2), 0leq pleq 1$, you'll assume that $p$ is more likely to be around $0.5$ compared to edge cases, such as $p=0,p=1$.
$endgroup$
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
add a comment |
$begingroup$
The notion of uniform prior as a prior with a constant density is not well-defined (or even meaningful) as it depends on both
- the dominating measure that determines the density function of the prior
- the parameterisation of the model for which the prior is constructed.
If either entry is modified, the density of the prior changes as well and stops being constant.
$endgroup$
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
add a comment |
$begingroup$
When the prior distribution $pi$, of the parameter $theta$ to be estimated is the Uniform distribution, i.e. $pi(theta)sim U(a,b)$, we refer to prior $pi$ as a uniform or uninformative prior. I'm not sure what's not to understand here except the basics of Bayesian inference and the Uniform distribution.
The best way to understand the uniform distribution is via a Monte Carlo sample of fair die rolls. The probability of every outcome $P(X=x_i),~ i=1,...,6$ is equal to $1/6$ and the histogram of the empirical distribution should be approximating a horizontal line (i.e. is uniform).
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "65"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f419529%2fcan-anyone-give-a-concrete-example-to-illustrate-what-is-an-uniform-prior%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Let's say you don't know the probability of head, $p$, of a coin. You decide to conduct an experiment to estimate what it is, via Bayesian analysis. It requires you to choose a prior, and in general you're free to choose one of the feasible ones. If you don't know or don't want to assume anything about this $p$, you can say that it is uniformly distributed in $[0,1]$, in which $f_P(p)=1, 0leq pleq 1$, $0$ otherwise. This is quite similar to saying that any $p$ value in $[0,1]$ is equally likely. This prior distribution is a uniform prior.
You can also choose other priors, that focus on different regions in $[0,1]$, for example, if you choose a prior like $f_P(p)=frac32(1-(1-2p)^2), 0leq pleq 1$, you'll assume that $p$ is more likely to be around $0.5$ compared to edge cases, such as $p=0,p=1$.
$endgroup$
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
add a comment |
$begingroup$
Let's say you don't know the probability of head, $p$, of a coin. You decide to conduct an experiment to estimate what it is, via Bayesian analysis. It requires you to choose a prior, and in general you're free to choose one of the feasible ones. If you don't know or don't want to assume anything about this $p$, you can say that it is uniformly distributed in $[0,1]$, in which $f_P(p)=1, 0leq pleq 1$, $0$ otherwise. This is quite similar to saying that any $p$ value in $[0,1]$ is equally likely. This prior distribution is a uniform prior.
You can also choose other priors, that focus on different regions in $[0,1]$, for example, if you choose a prior like $f_P(p)=frac32(1-(1-2p)^2), 0leq pleq 1$, you'll assume that $p$ is more likely to be around $0.5$ compared to edge cases, such as $p=0,p=1$.
$endgroup$
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
add a comment |
$begingroup$
Let's say you don't know the probability of head, $p$, of a coin. You decide to conduct an experiment to estimate what it is, via Bayesian analysis. It requires you to choose a prior, and in general you're free to choose one of the feasible ones. If you don't know or don't want to assume anything about this $p$, you can say that it is uniformly distributed in $[0,1]$, in which $f_P(p)=1, 0leq pleq 1$, $0$ otherwise. This is quite similar to saying that any $p$ value in $[0,1]$ is equally likely. This prior distribution is a uniform prior.
You can also choose other priors, that focus on different regions in $[0,1]$, for example, if you choose a prior like $f_P(p)=frac32(1-(1-2p)^2), 0leq pleq 1$, you'll assume that $p$ is more likely to be around $0.5$ compared to edge cases, such as $p=0,p=1$.
$endgroup$
Let's say you don't know the probability of head, $p$, of a coin. You decide to conduct an experiment to estimate what it is, via Bayesian analysis. It requires you to choose a prior, and in general you're free to choose one of the feasible ones. If you don't know or don't want to assume anything about this $p$, you can say that it is uniformly distributed in $[0,1]$, in which $f_P(p)=1, 0leq pleq 1$, $0$ otherwise. This is quite similar to saying that any $p$ value in $[0,1]$ is equally likely. This prior distribution is a uniform prior.
You can also choose other priors, that focus on different regions in $[0,1]$, for example, if you choose a prior like $f_P(p)=frac32(1-(1-2p)^2), 0leq pleq 1$, you'll assume that $p$ is more likely to be around $0.5$ compared to edge cases, such as $p=0,p=1$.
edited 8 hours ago
answered 8 hours ago
gunesgunes
12.2k1 gold badge5 silver badges22 bronze badges
12.2k1 gold badge5 silver badges22 bronze badges
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
add a comment |
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
$begingroup$
Thank you so much! Does the function $f_P(p)$ represent the random variables, Bernoulli random variable in this case?
$endgroup$
– czlsws
1 hour ago
add a comment |
$begingroup$
The notion of uniform prior as a prior with a constant density is not well-defined (or even meaningful) as it depends on both
- the dominating measure that determines the density function of the prior
- the parameterisation of the model for which the prior is constructed.
If either entry is modified, the density of the prior changes as well and stops being constant.
$endgroup$
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
add a comment |
$begingroup$
The notion of uniform prior as a prior with a constant density is not well-defined (or even meaningful) as it depends on both
- the dominating measure that determines the density function of the prior
- the parameterisation of the model for which the prior is constructed.
If either entry is modified, the density of the prior changes as well and stops being constant.
$endgroup$
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
add a comment |
$begingroup$
The notion of uniform prior as a prior with a constant density is not well-defined (or even meaningful) as it depends on both
- the dominating measure that determines the density function of the prior
- the parameterisation of the model for which the prior is constructed.
If either entry is modified, the density of the prior changes as well and stops being constant.
$endgroup$
The notion of uniform prior as a prior with a constant density is not well-defined (or even meaningful) as it depends on both
- the dominating measure that determines the density function of the prior
- the parameterisation of the model for which the prior is constructed.
If either entry is modified, the density of the prior changes as well and stops being constant.
edited 6 hours ago
answered 8 hours ago
Xi'anXi'an
61.8k8 gold badges99 silver badges377 bronze badges
61.8k8 gold badges99 silver badges377 bronze badges
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
add a comment |
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
2
2
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
$begingroup$
+1. Another way to express this, perhaps a little more forcibly, is that all (absolutely continuous) priors are uniform.
$endgroup$
– whuber♦
7 hours ago
add a comment |
$begingroup$
When the prior distribution $pi$, of the parameter $theta$ to be estimated is the Uniform distribution, i.e. $pi(theta)sim U(a,b)$, we refer to prior $pi$ as a uniform or uninformative prior. I'm not sure what's not to understand here except the basics of Bayesian inference and the Uniform distribution.
The best way to understand the uniform distribution is via a Monte Carlo sample of fair die rolls. The probability of every outcome $P(X=x_i),~ i=1,...,6$ is equal to $1/6$ and the histogram of the empirical distribution should be approximating a horizontal line (i.e. is uniform).
$endgroup$
add a comment |
$begingroup$
When the prior distribution $pi$, of the parameter $theta$ to be estimated is the Uniform distribution, i.e. $pi(theta)sim U(a,b)$, we refer to prior $pi$ as a uniform or uninformative prior. I'm not sure what's not to understand here except the basics of Bayesian inference and the Uniform distribution.
The best way to understand the uniform distribution is via a Monte Carlo sample of fair die rolls. The probability of every outcome $P(X=x_i),~ i=1,...,6$ is equal to $1/6$ and the histogram of the empirical distribution should be approximating a horizontal line (i.e. is uniform).
$endgroup$
add a comment |
$begingroup$
When the prior distribution $pi$, of the parameter $theta$ to be estimated is the Uniform distribution, i.e. $pi(theta)sim U(a,b)$, we refer to prior $pi$ as a uniform or uninformative prior. I'm not sure what's not to understand here except the basics of Bayesian inference and the Uniform distribution.
The best way to understand the uniform distribution is via a Monte Carlo sample of fair die rolls. The probability of every outcome $P(X=x_i),~ i=1,...,6$ is equal to $1/6$ and the histogram of the empirical distribution should be approximating a horizontal line (i.e. is uniform).
$endgroup$
When the prior distribution $pi$, of the parameter $theta$ to be estimated is the Uniform distribution, i.e. $pi(theta)sim U(a,b)$, we refer to prior $pi$ as a uniform or uninformative prior. I'm not sure what's not to understand here except the basics of Bayesian inference and the Uniform distribution.
The best way to understand the uniform distribution is via a Monte Carlo sample of fair die rolls. The probability of every outcome $P(X=x_i),~ i=1,...,6$ is equal to $1/6$ and the histogram of the empirical distribution should be approximating a horizontal line (i.e. is uniform).
answered 5 hours ago
DigioDigio
1,8738 silver badges17 bronze badges
1,8738 silver badges17 bronze badges
add a comment |
add a comment |
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f419529%2fcan-anyone-give-a-concrete-example-to-illustrate-what-is-an-uniform-prior%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
Possibly useful link.
$endgroup$
– BruceET
8 hours ago