What is joint estimation?Variance-gamma distribution: parameter estimationSimilarities and differences between regression and estimationCan someone explain the concept of nontrivial joint sufficient statistic for a pair of parameters?What location parameter is modelled by robust regression?What does it mean for a sample size to be “too low”, in terms of estimation vs inferrence?Logistic regression diagnostic plots in R
What kind of tools would be used to carve bone?
How to make "acts of patience" exciting?
Looking for PC graphics demo software from the early 90s called "Unreal"
'Cheddar goes "good" with burgers?' Can "go" be seen as a verb of the senses?
Should I withdraw my paper because the editor is delaying the report?
Compare items between two instances
An employee has low self-confidence, and is performing poorly. How can I help?
Creating chess engine, machine learning vs. traditional engine?
How are steel imports supposed to threaten US national security?
A Society Built Around Theft?
Why can I ping 10.0.0.0/8 addresses from a 192.168.1.0/24 subnet?
A demigod among men
How can I float a pin that otherwise should be low?
Is the Olympic running race fair?
Is data science mathematically interesting?
What term would be used for words that are borrowed from Japanese and used in other languages?
What are the different ways one can refer to the home in everyday French
70's/80's story about a high ranking prisoner of war on a prison planet
What is joint estimation?
Why is the intercept changing in a logistic regression when all predictors are standardized?
Adding elements to some sublists of unequal length
Why CMYK & PNG is not possible?
What damaging options does a lich have while in an anti-magic field?
Suspicious crontab entry
What is joint estimation?
Variance-gamma distribution: parameter estimationSimilarities and differences between regression and estimationCan someone explain the concept of nontrivial joint sufficient statistic for a pair of parameters?What location parameter is modelled by robust regression?What does it mean for a sample size to be “too low”, in terms of estimation vs inferrence?Logistic regression diagnostic plots in R
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty
margin-bottom:0;
$begingroup$
My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.
regression estimation
$endgroup$
add a comment
|
$begingroup$
My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.
regression estimation
$endgroup$
add a comment
|
$begingroup$
My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.
regression estimation
$endgroup$
My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.
regression estimation
regression estimation
asked 8 hours ago
Lost in regressionLost in regression
312 bronze badges
312 bronze badges
add a comment
|
add a comment
|
3 Answers
3
active
oldest
votes
$begingroup$
Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.
In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.
One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.
$endgroup$
add a comment
|
$begingroup$
In a statistical context, the term "joint estimation" could conceivably mean one of two things:
- The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or
- The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).
Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.
$endgroup$
add a comment
|
$begingroup$
Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.
Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.
The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.
In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for $beta_x,beta_y,alpha$. You have a large, complete data set of $x,y,z$.
In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.
As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.
In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.
Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.
I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.
https://www.youtube.com/watch?v=6GhSiM0frIk
https://www.youtube.com/watch?v=5SPm4TmYTX0
https://www.youtube.com/watch?v=b1GxZdFN6cY
and
https://www.youtube.com/watch?v=WdnP1gmb8Hw.
$endgroup$
add a comment
|
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "65"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/4.0/"u003ecc by-sa 4.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f429712%2fwhat-is-joint-estimation%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.
In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.
One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.
$endgroup$
add a comment
|
$begingroup$
Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.
In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.
One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.
$endgroup$
add a comment
|
$begingroup$
Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.
In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.
One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.
$endgroup$
Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.
In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.
One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.
edited 6 hours ago
answered 7 hours ago
AdamOAdamO
38.8k2 gold badges71 silver badges156 bronze badges
38.8k2 gold badges71 silver badges156 bronze badges
add a comment
|
add a comment
|
$begingroup$
In a statistical context, the term "joint estimation" could conceivably mean one of two things:
- The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or
- The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).
Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.
$endgroup$
add a comment
|
$begingroup$
In a statistical context, the term "joint estimation" could conceivably mean one of two things:
- The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or
- The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).
Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.
$endgroup$
add a comment
|
$begingroup$
In a statistical context, the term "joint estimation" could conceivably mean one of two things:
- The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or
- The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).
Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.
$endgroup$
In a statistical context, the term "joint estimation" could conceivably mean one of two things:
- The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or
- The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).
Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.
answered 5 hours ago
BenBen
38.2k2 gold badges50 silver badges167 bronze badges
38.2k2 gold badges50 silver badges167 bronze badges
add a comment
|
add a comment
|
$begingroup$
Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.
Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.
The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.
In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for $beta_x,beta_y,alpha$. You have a large, complete data set of $x,y,z$.
In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.
As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.
In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.
Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.
I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.
https://www.youtube.com/watch?v=6GhSiM0frIk
https://www.youtube.com/watch?v=5SPm4TmYTX0
https://www.youtube.com/watch?v=b1GxZdFN6cY
and
https://www.youtube.com/watch?v=WdnP1gmb8Hw.
$endgroup$
add a comment
|
$begingroup$
Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.
Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.
The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.
In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for $beta_x,beta_y,alpha$. You have a large, complete data set of $x,y,z$.
In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.
As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.
In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.
Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.
I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.
https://www.youtube.com/watch?v=6GhSiM0frIk
https://www.youtube.com/watch?v=5SPm4TmYTX0
https://www.youtube.com/watch?v=b1GxZdFN6cY
and
https://www.youtube.com/watch?v=WdnP1gmb8Hw.
$endgroup$
add a comment
|
$begingroup$
Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.
Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.
The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.
In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for $beta_x,beta_y,alpha$. You have a large, complete data set of $x,y,z$.
In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.
As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.
In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.
Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.
I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.
https://www.youtube.com/watch?v=6GhSiM0frIk
https://www.youtube.com/watch?v=5SPm4TmYTX0
https://www.youtube.com/watch?v=b1GxZdFN6cY
and
https://www.youtube.com/watch?v=WdnP1gmb8Hw.
$endgroup$
Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.
Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.
The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.
In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for $beta_x,beta_y,alpha$. You have a large, complete data set of $x,y,z$.
In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.
As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.
In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.
Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.
I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.
https://www.youtube.com/watch?v=6GhSiM0frIk
https://www.youtube.com/watch?v=5SPm4TmYTX0
https://www.youtube.com/watch?v=b1GxZdFN6cY
and
https://www.youtube.com/watch?v=WdnP1gmb8Hw.
answered 6 hours ago
Dave HarrisDave Harris
4,3905 silver badges17 bronze badges
4,3905 silver badges17 bronze badges
add a comment
|
add a comment
|
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f429712%2fwhat-is-joint-estimation%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown