Linear regression with negative estimated value for intercept
Clash Royale CLAN TAG#URR8PPP
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty margin-bottom:0;
up vote
4
down vote
favorite
Does a negative value of intercept suggest that the regression line provides poor fit to the data? why? and why not?
regression goodness-of-fit model-evaluation intercept
add a comment |Â
up vote
4
down vote
favorite
Does a negative value of intercept suggest that the regression line provides poor fit to the data? why? and why not?
regression goodness-of-fit model-evaluation intercept
13
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
1
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago
add a comment |Â
up vote
4
down vote
favorite
up vote
4
down vote
favorite
Does a negative value of intercept suggest that the regression line provides poor fit to the data? why? and why not?
regression goodness-of-fit model-evaluation intercept
Does a negative value of intercept suggest that the regression line provides poor fit to the data? why? and why not?
regression goodness-of-fit model-evaluation intercept
edited Sep 5 at 1:37
gung♦
103k34244508
103k34244508
asked Sep 4 at 5:02
M.Nazir
241
241
13
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
1
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago
add a comment |Â
13
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
1
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago
13
13
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
1
1
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago
add a comment |Â
4 Answers
4
active
oldest
votes
up vote
11
down vote
This an example of linear regression fit.
The intercept of this fit is negative and it fits well.
add a comment |Â
up vote
5
down vote
No, a negative value of intercept does not suggest that the regression line provides “poor fit to the dataâ€Â. Why not? Because the intercept is not a measure of fit for a regression. It is a value that represents one element of the model, not the quality of the fit.
The goodness of fit of a statistical model describes how well it fits a set of observations. Measures of goodness of fit typically summarize the discrepancy between observed values and the values expected under the model in question. WP:EN s.v. Goodness of fit
[...] y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system. WP:EN s.v. y-intercept
See KDG's answer for an illustration.
add a comment |Â
up vote
3
down vote
As others have said, intercept is not a measure of fit at all, so a negative intercept does not indicate a poor fit. But -- you asked if a negative intercept suggests a poor fit. In some contexts, the answer is clearly "yes". If $X$ and $Y$ are physical quantities such that $Y < 0$ makes no physical sense, and furthermore values of $X$ close to 0 are non-outliers which are part of your data set, then a linear model $Y = mX + b$ with $b < 0$ (by a significant amount) won't be a good model for your data. Note that such a model will have large residuals for those $X$ values close to 0, so the poorness of fit will show up in the standard measures.
But in other contexts, a negative intercept is unproblematic even if $Y < 0$ makes no sense. As an example, R's built-in sample dataset trees gives height, volume and girth of 31 black cherry trees. If you do a linear regression of volume as a function of girth (lm(formula = Volume ~ Girth, data = trees)
) you get a linear model with a negative intercept. The fit isn't bad (even though a quadratic model would be better) with $R^2 = 0.93$. You could use it to get reasonable predictions of the volume of such a tree in terms of its girth. In this case, the negative intercept simply indicates that it is unreasonable to extrapolate a relationship which (roughly) holds among mature trees to mere saplings. But - you don't need a negative intercept to tell you that such extrapolation is problematic. I wouldn't trust any physical model with $X$ near 0 unless the data itself involved such values.
add a comment |Â
up vote
1
down vote
The intercept value has no relation to goodness-of-fit.
To see if a fit-line is good fit for the data you need to calculate the distance between the fit-line and all the data points (as shown in @KDG's answer).
One metric for calculating this is Root Mean Square Error (RMSE).
add a comment |Â
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
11
down vote
This an example of linear regression fit.
The intercept of this fit is negative and it fits well.
add a comment |Â
up vote
11
down vote
This an example of linear regression fit.
The intercept of this fit is negative and it fits well.
add a comment |Â
up vote
11
down vote
up vote
11
down vote
This an example of linear regression fit.
The intercept of this fit is negative and it fits well.
This an example of linear regression fit.
The intercept of this fit is negative and it fits well.
edited Sep 4 at 10:37
Nick Cox
36.9k477124
36.9k477124
answered Sep 4 at 5:19
KDG
471113
471113
add a comment |Â
add a comment |Â
up vote
5
down vote
No, a negative value of intercept does not suggest that the regression line provides “poor fit to the dataâ€Â. Why not? Because the intercept is not a measure of fit for a regression. It is a value that represents one element of the model, not the quality of the fit.
The goodness of fit of a statistical model describes how well it fits a set of observations. Measures of goodness of fit typically summarize the discrepancy between observed values and the values expected under the model in question. WP:EN s.v. Goodness of fit
[...] y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system. WP:EN s.v. y-intercept
See KDG's answer for an illustration.
add a comment |Â
up vote
5
down vote
No, a negative value of intercept does not suggest that the regression line provides “poor fit to the dataâ€Â. Why not? Because the intercept is not a measure of fit for a regression. It is a value that represents one element of the model, not the quality of the fit.
The goodness of fit of a statistical model describes how well it fits a set of observations. Measures of goodness of fit typically summarize the discrepancy between observed values and the values expected under the model in question. WP:EN s.v. Goodness of fit
[...] y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system. WP:EN s.v. y-intercept
See KDG's answer for an illustration.
add a comment |Â
up vote
5
down vote
up vote
5
down vote
No, a negative value of intercept does not suggest that the regression line provides “poor fit to the dataâ€Â. Why not? Because the intercept is not a measure of fit for a regression. It is a value that represents one element of the model, not the quality of the fit.
The goodness of fit of a statistical model describes how well it fits a set of observations. Measures of goodness of fit typically summarize the discrepancy between observed values and the values expected under the model in question. WP:EN s.v. Goodness of fit
[...] y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system. WP:EN s.v. y-intercept
See KDG's answer for an illustration.
No, a negative value of intercept does not suggest that the regression line provides “poor fit to the dataâ€Â. Why not? Because the intercept is not a measure of fit for a regression. It is a value that represents one element of the model, not the quality of the fit.
The goodness of fit of a statistical model describes how well it fits a set of observations. Measures of goodness of fit typically summarize the discrepancy between observed values and the values expected under the model in question. WP:EN s.v. Goodness of fit
[...] y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system. WP:EN s.v. y-intercept
See KDG's answer for an illustration.
edited Sep 4 at 9:54
answered Sep 4 at 9:25
marianoju
514
514
add a comment |Â
add a comment |Â
up vote
3
down vote
As others have said, intercept is not a measure of fit at all, so a negative intercept does not indicate a poor fit. But -- you asked if a negative intercept suggests a poor fit. In some contexts, the answer is clearly "yes". If $X$ and $Y$ are physical quantities such that $Y < 0$ makes no physical sense, and furthermore values of $X$ close to 0 are non-outliers which are part of your data set, then a linear model $Y = mX + b$ with $b < 0$ (by a significant amount) won't be a good model for your data. Note that such a model will have large residuals for those $X$ values close to 0, so the poorness of fit will show up in the standard measures.
But in other contexts, a negative intercept is unproblematic even if $Y < 0$ makes no sense. As an example, R's built-in sample dataset trees gives height, volume and girth of 31 black cherry trees. If you do a linear regression of volume as a function of girth (lm(formula = Volume ~ Girth, data = trees)
) you get a linear model with a negative intercept. The fit isn't bad (even though a quadratic model would be better) with $R^2 = 0.93$. You could use it to get reasonable predictions of the volume of such a tree in terms of its girth. In this case, the negative intercept simply indicates that it is unreasonable to extrapolate a relationship which (roughly) holds among mature trees to mere saplings. But - you don't need a negative intercept to tell you that such extrapolation is problematic. I wouldn't trust any physical model with $X$ near 0 unless the data itself involved such values.
add a comment |Â
up vote
3
down vote
As others have said, intercept is not a measure of fit at all, so a negative intercept does not indicate a poor fit. But -- you asked if a negative intercept suggests a poor fit. In some contexts, the answer is clearly "yes". If $X$ and $Y$ are physical quantities such that $Y < 0$ makes no physical sense, and furthermore values of $X$ close to 0 are non-outliers which are part of your data set, then a linear model $Y = mX + b$ with $b < 0$ (by a significant amount) won't be a good model for your data. Note that such a model will have large residuals for those $X$ values close to 0, so the poorness of fit will show up in the standard measures.
But in other contexts, a negative intercept is unproblematic even if $Y < 0$ makes no sense. As an example, R's built-in sample dataset trees gives height, volume and girth of 31 black cherry trees. If you do a linear regression of volume as a function of girth (lm(formula = Volume ~ Girth, data = trees)
) you get a linear model with a negative intercept. The fit isn't bad (even though a quadratic model would be better) with $R^2 = 0.93$. You could use it to get reasonable predictions of the volume of such a tree in terms of its girth. In this case, the negative intercept simply indicates that it is unreasonable to extrapolate a relationship which (roughly) holds among mature trees to mere saplings. But - you don't need a negative intercept to tell you that such extrapolation is problematic. I wouldn't trust any physical model with $X$ near 0 unless the data itself involved such values.
add a comment |Â
up vote
3
down vote
up vote
3
down vote
As others have said, intercept is not a measure of fit at all, so a negative intercept does not indicate a poor fit. But -- you asked if a negative intercept suggests a poor fit. In some contexts, the answer is clearly "yes". If $X$ and $Y$ are physical quantities such that $Y < 0$ makes no physical sense, and furthermore values of $X$ close to 0 are non-outliers which are part of your data set, then a linear model $Y = mX + b$ with $b < 0$ (by a significant amount) won't be a good model for your data. Note that such a model will have large residuals for those $X$ values close to 0, so the poorness of fit will show up in the standard measures.
But in other contexts, a negative intercept is unproblematic even if $Y < 0$ makes no sense. As an example, R's built-in sample dataset trees gives height, volume and girth of 31 black cherry trees. If you do a linear regression of volume as a function of girth (lm(formula = Volume ~ Girth, data = trees)
) you get a linear model with a negative intercept. The fit isn't bad (even though a quadratic model would be better) with $R^2 = 0.93$. You could use it to get reasonable predictions of the volume of such a tree in terms of its girth. In this case, the negative intercept simply indicates that it is unreasonable to extrapolate a relationship which (roughly) holds among mature trees to mere saplings. But - you don't need a negative intercept to tell you that such extrapolation is problematic. I wouldn't trust any physical model with $X$ near 0 unless the data itself involved such values.
As others have said, intercept is not a measure of fit at all, so a negative intercept does not indicate a poor fit. But -- you asked if a negative intercept suggests a poor fit. In some contexts, the answer is clearly "yes". If $X$ and $Y$ are physical quantities such that $Y < 0$ makes no physical sense, and furthermore values of $X$ close to 0 are non-outliers which are part of your data set, then a linear model $Y = mX + b$ with $b < 0$ (by a significant amount) won't be a good model for your data. Note that such a model will have large residuals for those $X$ values close to 0, so the poorness of fit will show up in the standard measures.
But in other contexts, a negative intercept is unproblematic even if $Y < 0$ makes no sense. As an example, R's built-in sample dataset trees gives height, volume and girth of 31 black cherry trees. If you do a linear regression of volume as a function of girth (lm(formula = Volume ~ Girth, data = trees)
) you get a linear model with a negative intercept. The fit isn't bad (even though a quadratic model would be better) with $R^2 = 0.93$. You could use it to get reasonable predictions of the volume of such a tree in terms of its girth. In this case, the negative intercept simply indicates that it is unreasonable to extrapolate a relationship which (roughly) holds among mature trees to mere saplings. But - you don't need a negative intercept to tell you that such extrapolation is problematic. I wouldn't trust any physical model with $X$ near 0 unless the data itself involved such values.
edited Sep 4 at 12:00
answered Sep 4 at 11:54


John Coleman
22316
22316
add a comment |Â
add a comment |Â
up vote
1
down vote
The intercept value has no relation to goodness-of-fit.
To see if a fit-line is good fit for the data you need to calculate the distance between the fit-line and all the data points (as shown in @KDG's answer).
One metric for calculating this is Root Mean Square Error (RMSE).
add a comment |Â
up vote
1
down vote
The intercept value has no relation to goodness-of-fit.
To see if a fit-line is good fit for the data you need to calculate the distance between the fit-line and all the data points (as shown in @KDG's answer).
One metric for calculating this is Root Mean Square Error (RMSE).
add a comment |Â
up vote
1
down vote
up vote
1
down vote
The intercept value has no relation to goodness-of-fit.
To see if a fit-line is good fit for the data you need to calculate the distance between the fit-line and all the data points (as shown in @KDG's answer).
One metric for calculating this is Root Mean Square Error (RMSE).
The intercept value has no relation to goodness-of-fit.
To see if a fit-line is good fit for the data you need to calculate the distance between the fit-line and all the data points (as shown in @KDG's answer).
One metric for calculating this is Root Mean Square Error (RMSE).
answered Sep 4 at 10:45


Paresh
111
111
add a comment |Â
add a comment |Â
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f365299%2flinear-regression-with-negative-estimated-value-for-intercept%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
13
The phrasing of your question implies that you have some reason to think the premise may be true. Why? What prompts such a question? (I ask because I suspect there may be a deeper issue laying beneath this one that should be cleared up -- possibly indicating the unsuitability of a linear regression model for whatever you're thinking about)
– Glen_b♦
Sep 4 at 6:03
1
If you don't like negative intercepts, just negate the response variable. That will not change the linearity (that is, goodness of fit) of any model but will produce a positive intercept.
– whuber♦
Sep 4 at 13:56
Unless it's a variable where a negative value would be meaningless.
– smci
Sep 5 at 0:40
Have you have found the answers helpful? Have you considered upvoting and accepting an answer? If you find that no one provides a satisfactory answer to your question you might want to consider editing your question(s) to provide more, clearer information as to the nature of the problem. If you find that your question is not getting good answers and you want to motivate answerers, you might want to consider starting a bounty for this question.
– marianoju
5 hours ago