Why is the sum of probabilities in a continuous uniform distribution not infinity?
Clash Royale CLAN TAG#URR8PPP
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty margin-bottom:0;
up vote
9
down vote
favorite
The probability density function of a uniform distribution (continuous) is shown above. The area under the curve is 1 - which makes sense since the sum of all the probabilities in a probability distribution is 1.
Formally, the above probability function (f(x)) can be defined as
1/(b-a) for x in [a,b]
and 0 otherwise
Consider that I have to choose a real number between a (say, 2) and b (say, 6). This makes the uniform probability = 0.25. However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity? What am I overlooking?
Is f(x) not the probability of the number x occurring?
probability distributions uniform
add a comment |Â
up vote
9
down vote
favorite
The probability density function of a uniform distribution (continuous) is shown above. The area under the curve is 1 - which makes sense since the sum of all the probabilities in a probability distribution is 1.
Formally, the above probability function (f(x)) can be defined as
1/(b-a) for x in [a,b]
and 0 otherwise
Consider that I have to choose a real number between a (say, 2) and b (say, 6). This makes the uniform probability = 0.25. However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity? What am I overlooking?
Is f(x) not the probability of the number x occurring?
probability distributions uniform
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29
add a comment |Â
up vote
9
down vote
favorite
up vote
9
down vote
favorite
The probability density function of a uniform distribution (continuous) is shown above. The area under the curve is 1 - which makes sense since the sum of all the probabilities in a probability distribution is 1.
Formally, the above probability function (f(x)) can be defined as
1/(b-a) for x in [a,b]
and 0 otherwise
Consider that I have to choose a real number between a (say, 2) and b (say, 6). This makes the uniform probability = 0.25. However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity? What am I overlooking?
Is f(x) not the probability of the number x occurring?
probability distributions uniform
The probability density function of a uniform distribution (continuous) is shown above. The area under the curve is 1 - which makes sense since the sum of all the probabilities in a probability distribution is 1.
Formally, the above probability function (f(x)) can be defined as
1/(b-a) for x in [a,b]
and 0 otherwise
Consider that I have to choose a real number between a (say, 2) and b (say, 6). This makes the uniform probability = 0.25. However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity? What am I overlooking?
Is f(x) not the probability of the number x occurring?
probability distributions uniform
asked Aug 28 at 3:18
Rahul Saha
15517
15517
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29
add a comment |Â
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29
add a comment |Â
5 Answers
5
active
oldest
votes
up vote
18
down vote
accepted
$f(x)$ describes the probability density rather than a probability mass in your example. In general, for continuous distributions the events—the things we get probabilities for—are ranges of values, such as for the area under the curve from $a$ to $a+.1$, or from $a$ to $b$ (although such ranges need not be contiguous). For continuous distributions, the probability of any single value occurring is generally 0.
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
 |Â
show 1 more comment
up vote
11
down vote
Because each term in the summation is weighted by the infinitesimal d$x$. The importance of this is probably most easily understood by carefully walking through a very basic example.
Consider using Riemann summation to compute the area under the following rectangular region (a rectangle was chosen to remove the approximation aspect of Riemann summation, which is not the focus here):]
We can compute the area using 2 subregions, or by using 4 subregions. In the case of the 2 subregions (denoted $A_i$), the areas are given by $$A_1=A_2=5times 2 = 10$$ whereas in the case of 4 subregions (denoted $B_i$), the areas are given by $$B_1=B_2=B_3=B_4=5times 1 = 5$$ The total area in both cases correspond to $$sum_i=1^2 A_i = sum_i=1^4B_i = 20$$
Now, this is all fairly obvious, but it raises a subtly important question which is: why do these two answers agree? Intuitively it should be clear that it works because we've reduced the width of the second set of subregions. We could consider doing the same thing with 8 subregions each with a width of $0.5$, and again with 16... and we could continue this process until we have an infinite number of subregions, each with a tiny width of d$x$. As long as everything is always correctly weighted, the answers should always agree. Without the correct weighting, the summation would indeed simply be $infty$.
This is why I always make sure to point out to students that an integral is not simply the symbol $int$, but the pair of symbols $int text dx$.
add a comment |Â
up vote
4
down vote
You're interpreting the probability distribution the wrong way - it is an infinite number of infinitely divided probabilities, so you can't say that "the probability of drawing the value 0.5 from a (0, 1) uniform distribution" because that probability is zero - there are an infinite number of possible values you could get, and all of them are equally likely, so clearly the probability of any individual outcome is $frac1infty = 0$[1].
Instead, you can look at the probability for a range of outcomes, and measure that using areas (and hence integrals). For example, if you draw from the (0, 1) uniform distribution (with pdf $f(x) = 1$ for $x in left[0, 1right]$ and $f(x) = 0$ otherwise), then the probability that your result lies between $0.2$ and $0.3$ is
$int_0.2^0.3 f(x) dx = int_0.2^0.3 1 dx = left[x right]_0.2^0.3 = 0.3 - 0.2 = 0.1$
i.e. you have a 10% chance of getting a result in that range.
[1]Sorry for all the people having heart attacks at my over-simplification of the calculation.
add a comment |Â
up vote
0
down vote
In general your reasoning fails in this assumption:
However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity?
It's a mathematical problem, known since the Zeno of Elea Paradoxes.
Two of his claims were that
- An arrow can never reach its target
- Achilles will never overtake a turtle
Both of them were based on the claim that you can build an infinite sequence of positive numbers (in the former case by saying that an arrow has to fly infinitely times half of the remaining way to the target, in the latter by saying that Achilles has to reach position where the turtle was previously, and in the meantime the turtle moves to a new position that becomes our next reference base point).
Fast forward, this led to a discovery of infinite sums.
So in general sum of infinite many positive numbers does not necessarily have to be infinite; however, it may not be infinite only if (an extreme oversimplification, sorry about that) almost all of the numbers in the sequence are very close to 0, regardless how close to zero you request them to be.
Infinity plays even more tricks. The order in which you add elements of the sequence is important too and might lead to a situation that reordering gives different results!
Explore a bit more about paradoxes of infinity. You might be astonished.
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
add a comment |Â
up vote
0
down vote
$f(x)$ describes the probability density, and has the unit $fracpx$. Hence for a given x you get $f(x) = frac1b-a$ in $fracpx$ units, and not p, as you are looking for. If you want p, you need the distribution function for a given range, that is the probability p of x being within a and b.
Hope this makes sense.
add a comment |Â
5 Answers
5
active
oldest
votes
5 Answers
5
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
18
down vote
accepted
$f(x)$ describes the probability density rather than a probability mass in your example. In general, for continuous distributions the events—the things we get probabilities for—are ranges of values, such as for the area under the curve from $a$ to $a+.1$, or from $a$ to $b$ (although such ranges need not be contiguous). For continuous distributions, the probability of any single value occurring is generally 0.
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
 |Â
show 1 more comment
up vote
18
down vote
accepted
$f(x)$ describes the probability density rather than a probability mass in your example. In general, for continuous distributions the events—the things we get probabilities for—are ranges of values, such as for the area under the curve from $a$ to $a+.1$, or from $a$ to $b$ (although such ranges need not be contiguous). For continuous distributions, the probability of any single value occurring is generally 0.
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
 |Â
show 1 more comment
up vote
18
down vote
accepted
up vote
18
down vote
accepted
$f(x)$ describes the probability density rather than a probability mass in your example. In general, for continuous distributions the events—the things we get probabilities for—are ranges of values, such as for the area under the curve from $a$ to $a+.1$, or from $a$ to $b$ (although such ranges need not be contiguous). For continuous distributions, the probability of any single value occurring is generally 0.
$f(x)$ describes the probability density rather than a probability mass in your example. In general, for continuous distributions the events—the things we get probabilities for—are ranges of values, such as for the area under the curve from $a$ to $a+.1$, or from $a$ to $b$ (although such ranges need not be contiguous). For continuous distributions, the probability of any single value occurring is generally 0.
edited Aug 28 at 14:21
answered Aug 28 at 3:32
Alexis
15.3k34493
15.3k34493
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
 |Â
show 1 more comment
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
Is there a more technically accurate way to say what you're trying to say? I'm worried the "range" thing will throw people off, considering continuous distributions can have Dirac deltas...
– Mehrdad
Aug 28 at 9:03
3
3
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
@Mehrdad: The dirac delta does not have a continuous distribution. The proper way of assigning probabilities would be via $P(A) = int_A 1dF$.
– Alex R.
Aug 28 at 10:39
1
1
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@AlexR.: Oof, I assumed by "continuous distribution" you just meant a distribution over a continuous domain, since that's what people refer to when they say the Dirac delta is the continuous analog of the Kronecker delta. Thanks for clarifying.
– Mehrdad
Aug 28 at 11:24
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad I was thinking precisely of Dirac's delta, but I hope you will notice the term "in general", and also the apparent level of statistical literacy of the OP.
– Alexis
Aug 28 at 13:57
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
@Mehrdad The technical formulation of a random variable is in terms of a measure: there is a function from the power set of the event space to the interval [0,1]. A probability density function can be used as a measure (the measure of a set is simply the integral of the PDF over that set), but there are measures, such as the Dirac delta (a set has measure 1 if it contains $x_0$, and is zero otherwise) that are, strictly speaking, not function in the traditional sense.
– Acccumulation
Aug 29 at 14:51
 |Â
show 1 more comment
up vote
11
down vote
Because each term in the summation is weighted by the infinitesimal d$x$. The importance of this is probably most easily understood by carefully walking through a very basic example.
Consider using Riemann summation to compute the area under the following rectangular region (a rectangle was chosen to remove the approximation aspect of Riemann summation, which is not the focus here):]
We can compute the area using 2 subregions, or by using 4 subregions. In the case of the 2 subregions (denoted $A_i$), the areas are given by $$A_1=A_2=5times 2 = 10$$ whereas in the case of 4 subregions (denoted $B_i$), the areas are given by $$B_1=B_2=B_3=B_4=5times 1 = 5$$ The total area in both cases correspond to $$sum_i=1^2 A_i = sum_i=1^4B_i = 20$$
Now, this is all fairly obvious, but it raises a subtly important question which is: why do these two answers agree? Intuitively it should be clear that it works because we've reduced the width of the second set of subregions. We could consider doing the same thing with 8 subregions each with a width of $0.5$, and again with 16... and we could continue this process until we have an infinite number of subregions, each with a tiny width of d$x$. As long as everything is always correctly weighted, the answers should always agree. Without the correct weighting, the summation would indeed simply be $infty$.
This is why I always make sure to point out to students that an integral is not simply the symbol $int$, but the pair of symbols $int text dx$.
add a comment |Â
up vote
11
down vote
Because each term in the summation is weighted by the infinitesimal d$x$. The importance of this is probably most easily understood by carefully walking through a very basic example.
Consider using Riemann summation to compute the area under the following rectangular region (a rectangle was chosen to remove the approximation aspect of Riemann summation, which is not the focus here):]
We can compute the area using 2 subregions, or by using 4 subregions. In the case of the 2 subregions (denoted $A_i$), the areas are given by $$A_1=A_2=5times 2 = 10$$ whereas in the case of 4 subregions (denoted $B_i$), the areas are given by $$B_1=B_2=B_3=B_4=5times 1 = 5$$ The total area in both cases correspond to $$sum_i=1^2 A_i = sum_i=1^4B_i = 20$$
Now, this is all fairly obvious, but it raises a subtly important question which is: why do these two answers agree? Intuitively it should be clear that it works because we've reduced the width of the second set of subregions. We could consider doing the same thing with 8 subregions each with a width of $0.5$, and again with 16... and we could continue this process until we have an infinite number of subregions, each with a tiny width of d$x$. As long as everything is always correctly weighted, the answers should always agree. Without the correct weighting, the summation would indeed simply be $infty$.
This is why I always make sure to point out to students that an integral is not simply the symbol $int$, but the pair of symbols $int text dx$.
add a comment |Â
up vote
11
down vote
up vote
11
down vote
Because each term in the summation is weighted by the infinitesimal d$x$. The importance of this is probably most easily understood by carefully walking through a very basic example.
Consider using Riemann summation to compute the area under the following rectangular region (a rectangle was chosen to remove the approximation aspect of Riemann summation, which is not the focus here):]
We can compute the area using 2 subregions, or by using 4 subregions. In the case of the 2 subregions (denoted $A_i$), the areas are given by $$A_1=A_2=5times 2 = 10$$ whereas in the case of 4 subregions (denoted $B_i$), the areas are given by $$B_1=B_2=B_3=B_4=5times 1 = 5$$ The total area in both cases correspond to $$sum_i=1^2 A_i = sum_i=1^4B_i = 20$$
Now, this is all fairly obvious, but it raises a subtly important question which is: why do these two answers agree? Intuitively it should be clear that it works because we've reduced the width of the second set of subregions. We could consider doing the same thing with 8 subregions each with a width of $0.5$, and again with 16... and we could continue this process until we have an infinite number of subregions, each with a tiny width of d$x$. As long as everything is always correctly weighted, the answers should always agree. Without the correct weighting, the summation would indeed simply be $infty$.
This is why I always make sure to point out to students that an integral is not simply the symbol $int$, but the pair of symbols $int text dx$.
Because each term in the summation is weighted by the infinitesimal d$x$. The importance of this is probably most easily understood by carefully walking through a very basic example.
Consider using Riemann summation to compute the area under the following rectangular region (a rectangle was chosen to remove the approximation aspect of Riemann summation, which is not the focus here):]
We can compute the area using 2 subregions, or by using 4 subregions. In the case of the 2 subregions (denoted $A_i$), the areas are given by $$A_1=A_2=5times 2 = 10$$ whereas in the case of 4 subregions (denoted $B_i$), the areas are given by $$B_1=B_2=B_3=B_4=5times 1 = 5$$ The total area in both cases correspond to $$sum_i=1^2 A_i = sum_i=1^4B_i = 20$$
Now, this is all fairly obvious, but it raises a subtly important question which is: why do these two answers agree? Intuitively it should be clear that it works because we've reduced the width of the second set of subregions. We could consider doing the same thing with 8 subregions each with a width of $0.5$, and again with 16... and we could continue this process until we have an infinite number of subregions, each with a tiny width of d$x$. As long as everything is always correctly weighted, the answers should always agree. Without the correct weighting, the summation would indeed simply be $infty$.
This is why I always make sure to point out to students that an integral is not simply the symbol $int$, but the pair of symbols $int text dx$.
edited Aug 28 at 6:11
answered Aug 28 at 6:02
Zxv
1113
1113
add a comment |Â
add a comment |Â
up vote
4
down vote
You're interpreting the probability distribution the wrong way - it is an infinite number of infinitely divided probabilities, so you can't say that "the probability of drawing the value 0.5 from a (0, 1) uniform distribution" because that probability is zero - there are an infinite number of possible values you could get, and all of them are equally likely, so clearly the probability of any individual outcome is $frac1infty = 0$[1].
Instead, you can look at the probability for a range of outcomes, and measure that using areas (and hence integrals). For example, if you draw from the (0, 1) uniform distribution (with pdf $f(x) = 1$ for $x in left[0, 1right]$ and $f(x) = 0$ otherwise), then the probability that your result lies between $0.2$ and $0.3$ is
$int_0.2^0.3 f(x) dx = int_0.2^0.3 1 dx = left[x right]_0.2^0.3 = 0.3 - 0.2 = 0.1$
i.e. you have a 10% chance of getting a result in that range.
[1]Sorry for all the people having heart attacks at my over-simplification of the calculation.
add a comment |Â
up vote
4
down vote
You're interpreting the probability distribution the wrong way - it is an infinite number of infinitely divided probabilities, so you can't say that "the probability of drawing the value 0.5 from a (0, 1) uniform distribution" because that probability is zero - there are an infinite number of possible values you could get, and all of them are equally likely, so clearly the probability of any individual outcome is $frac1infty = 0$[1].
Instead, you can look at the probability for a range of outcomes, and measure that using areas (and hence integrals). For example, if you draw from the (0, 1) uniform distribution (with pdf $f(x) = 1$ for $x in left[0, 1right]$ and $f(x) = 0$ otherwise), then the probability that your result lies between $0.2$ and $0.3$ is
$int_0.2^0.3 f(x) dx = int_0.2^0.3 1 dx = left[x right]_0.2^0.3 = 0.3 - 0.2 = 0.1$
i.e. you have a 10% chance of getting a result in that range.
[1]Sorry for all the people having heart attacks at my over-simplification of the calculation.
add a comment |Â
up vote
4
down vote
up vote
4
down vote
You're interpreting the probability distribution the wrong way - it is an infinite number of infinitely divided probabilities, so you can't say that "the probability of drawing the value 0.5 from a (0, 1) uniform distribution" because that probability is zero - there are an infinite number of possible values you could get, and all of them are equally likely, so clearly the probability of any individual outcome is $frac1infty = 0$[1].
Instead, you can look at the probability for a range of outcomes, and measure that using areas (and hence integrals). For example, if you draw from the (0, 1) uniform distribution (with pdf $f(x) = 1$ for $x in left[0, 1right]$ and $f(x) = 0$ otherwise), then the probability that your result lies between $0.2$ and $0.3$ is
$int_0.2^0.3 f(x) dx = int_0.2^0.3 1 dx = left[x right]_0.2^0.3 = 0.3 - 0.2 = 0.1$
i.e. you have a 10% chance of getting a result in that range.
[1]Sorry for all the people having heart attacks at my over-simplification of the calculation.
You're interpreting the probability distribution the wrong way - it is an infinite number of infinitely divided probabilities, so you can't say that "the probability of drawing the value 0.5 from a (0, 1) uniform distribution" because that probability is zero - there are an infinite number of possible values you could get, and all of them are equally likely, so clearly the probability of any individual outcome is $frac1infty = 0$[1].
Instead, you can look at the probability for a range of outcomes, and measure that using areas (and hence integrals). For example, if you draw from the (0, 1) uniform distribution (with pdf $f(x) = 1$ for $x in left[0, 1right]$ and $f(x) = 0$ otherwise), then the probability that your result lies between $0.2$ and $0.3$ is
$int_0.2^0.3 f(x) dx = int_0.2^0.3 1 dx = left[x right]_0.2^0.3 = 0.3 - 0.2 = 0.1$
i.e. you have a 10% chance of getting a result in that range.
[1]Sorry for all the people having heart attacks at my over-simplification of the calculation.
answered Aug 28 at 5:55
ConMan
1713
1713
add a comment |Â
add a comment |Â
up vote
0
down vote
In general your reasoning fails in this assumption:
However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity?
It's a mathematical problem, known since the Zeno of Elea Paradoxes.
Two of his claims were that
- An arrow can never reach its target
- Achilles will never overtake a turtle
Both of them were based on the claim that you can build an infinite sequence of positive numbers (in the former case by saying that an arrow has to fly infinitely times half of the remaining way to the target, in the latter by saying that Achilles has to reach position where the turtle was previously, and in the meantime the turtle moves to a new position that becomes our next reference base point).
Fast forward, this led to a discovery of infinite sums.
So in general sum of infinite many positive numbers does not necessarily have to be infinite; however, it may not be infinite only if (an extreme oversimplification, sorry about that) almost all of the numbers in the sequence are very close to 0, regardless how close to zero you request them to be.
Infinity plays even more tricks. The order in which you add elements of the sequence is important too and might lead to a situation that reordering gives different results!
Explore a bit more about paradoxes of infinity. You might be astonished.
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
add a comment |Â
up vote
0
down vote
In general your reasoning fails in this assumption:
However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity?
It's a mathematical problem, known since the Zeno of Elea Paradoxes.
Two of his claims were that
- An arrow can never reach its target
- Achilles will never overtake a turtle
Both of them were based on the claim that you can build an infinite sequence of positive numbers (in the former case by saying that an arrow has to fly infinitely times half of the remaining way to the target, in the latter by saying that Achilles has to reach position where the turtle was previously, and in the meantime the turtle moves to a new position that becomes our next reference base point).
Fast forward, this led to a discovery of infinite sums.
So in general sum of infinite many positive numbers does not necessarily have to be infinite; however, it may not be infinite only if (an extreme oversimplification, sorry about that) almost all of the numbers in the sequence are very close to 0, regardless how close to zero you request them to be.
Infinity plays even more tricks. The order in which you add elements of the sequence is important too and might lead to a situation that reordering gives different results!
Explore a bit more about paradoxes of infinity. You might be astonished.
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
add a comment |Â
up vote
0
down vote
up vote
0
down vote
In general your reasoning fails in this assumption:
However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity?
It's a mathematical problem, known since the Zeno of Elea Paradoxes.
Two of his claims were that
- An arrow can never reach its target
- Achilles will never overtake a turtle
Both of them were based on the claim that you can build an infinite sequence of positive numbers (in the former case by saying that an arrow has to fly infinitely times half of the remaining way to the target, in the latter by saying that Achilles has to reach position where the turtle was previously, and in the meantime the turtle moves to a new position that becomes our next reference base point).
Fast forward, this led to a discovery of infinite sums.
So in general sum of infinite many positive numbers does not necessarily have to be infinite; however, it may not be infinite only if (an extreme oversimplification, sorry about that) almost all of the numbers in the sequence are very close to 0, regardless how close to zero you request them to be.
Infinity plays even more tricks. The order in which you add elements of the sequence is important too and might lead to a situation that reordering gives different results!
Explore a bit more about paradoxes of infinity. You might be astonished.
In general your reasoning fails in this assumption:
However, since there are an infinite number of numbers in that interval, shouldn't the sum of all the probabilities sum up to infinity?
It's a mathematical problem, known since the Zeno of Elea Paradoxes.
Two of his claims were that
- An arrow can never reach its target
- Achilles will never overtake a turtle
Both of them were based on the claim that you can build an infinite sequence of positive numbers (in the former case by saying that an arrow has to fly infinitely times half of the remaining way to the target, in the latter by saying that Achilles has to reach position where the turtle was previously, and in the meantime the turtle moves to a new position that becomes our next reference base point).
Fast forward, this led to a discovery of infinite sums.
So in general sum of infinite many positive numbers does not necessarily have to be infinite; however, it may not be infinite only if (an extreme oversimplification, sorry about that) almost all of the numbers in the sequence are very close to 0, regardless how close to zero you request them to be.
Infinity plays even more tricks. The order in which you add elements of the sequence is important too and might lead to a situation that reordering gives different results!
Explore a bit more about paradoxes of infinity. You might be astonished.
edited Aug 28 at 12:03
Nick Cox
36.9k477124
36.9k477124
answered Aug 28 at 9:12
Ister
1011
1011
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
add a comment |Â
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
I don't see a way to interpret the question such that OP thinks of countable sums.
– JiK
Aug 28 at 15:13
add a comment |Â
up vote
0
down vote
$f(x)$ describes the probability density, and has the unit $fracpx$. Hence for a given x you get $f(x) = frac1b-a$ in $fracpx$ units, and not p, as you are looking for. If you want p, you need the distribution function for a given range, that is the probability p of x being within a and b.
Hope this makes sense.
add a comment |Â
up vote
0
down vote
$f(x)$ describes the probability density, and has the unit $fracpx$. Hence for a given x you get $f(x) = frac1b-a$ in $fracpx$ units, and not p, as you are looking for. If you want p, you need the distribution function for a given range, that is the probability p of x being within a and b.
Hope this makes sense.
add a comment |Â
up vote
0
down vote
up vote
0
down vote
$f(x)$ describes the probability density, and has the unit $fracpx$. Hence for a given x you get $f(x) = frac1b-a$ in $fracpx$ units, and not p, as you are looking for. If you want p, you need the distribution function for a given range, that is the probability p of x being within a and b.
Hope this makes sense.
$f(x)$ describes the probability density, and has the unit $fracpx$. Hence for a given x you get $f(x) = frac1b-a$ in $fracpx$ units, and not p, as you are looking for. If you want p, you need the distribution function for a given range, that is the probability p of x being within a and b.
Hope this makes sense.
edited Aug 28 at 14:04
Nick Cox
36.9k477124
36.9k477124
answered Aug 28 at 12:37
user3719750
11
11
add a comment |Â
add a comment |Â
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f364278%2fwhy-is-the-sum-of-probabilities-in-a-continuous-uniform-distribution-not-infinit%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
also: math.stackexchange.com/questions/2885278/… ; stats.stackexchange.com/questions/199280/…
– Ben Bolker
Aug 28 at 12:36
$f(x)$ is not a probability function—it is a probability density function. That is, it doesn't give you the probability for $x$ being a certain number, but the probability density, or the probability per unit length along the x-axis. You use integration to get the total probability for this type of function—not summation.
– HelloGoodbye
Aug 28 at 16:29