who's up for a nice meaty probability problem?
+4
Hellsangel
Petrichor
bw
MaxEntropy_Man
8 posters
Page 1 of 1
who's up for a nice meaty probability problem?
will post if there is an interest.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
let me just post the link directly.
to those who attempt: don't post methodology for a couple of weeks until they post the solution. just post your answer so we can compare answers. i sent them my solution today. i hope tiger sons and lion daughters are enthused enough to send solutions too.
to those who attempt: don't post methodology for a couple of weeks until they post the solution. just post your answer so we can compare answers. i sent them my solution today. i hope tiger sons and lion daughters are enthused enough to send solutions too.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Petrichor- Posts : 1725
Join date : 2012-04-10
Petrichor- Posts : 1725
Join date : 2012-04-10
Re: who's up for a nice meaty probability problem?
atcg wrote:and yours?
i'll wait for a few more responses. mine was higher than yours.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
1/216 =0.004629
Hellsangel- Posts : 14721
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
here is a helpful way to think about problems like this. think up and solve a related but simpler problem. in this case, a 1-d random walk played with a two-sided die with 1 & 2 marked on the two sides of the die (or alternately flip a coin). if you roll the die and get 1 you move one unit rightwards, i.e. +x, and if you roll the die and get 2, you go one unit leftwards, i.e. -x. what is the probability that you are back where you started after two rolls of the die? now generalize the result to 3-d, a six faced die, and six rolls of the die.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Hellsangel wrote:1/216 =0.004629
HA - not the answer i got.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
MaxEntropy_Man wrote:and if you roll the die and get 2, you go one unit leftwards, i.e. -x.
Should be -2 leftwards....the word used is "respectively"
Oops...you are right...withdrawing the answer.
Petrichor- Posts : 1725
Join date : 2012-04-10
Re: who's up for a nice meaty probability problem?
atcg wrote:MaxEntropy_Man wrote:and if you roll the die and get 2, you go one unit leftwards, i.e. -x.
Should be -2 leftwards....the word used is "respectively"
atcg: i don't think so. i think the "respectively" is just to keep track of labels. read the line that says the walker always jumps to one of its six nearest neighbors, i.e. the jump is always one lattice distance and the matchups are:
1 = +1 (x-direction), 2 = -1 (x-direction), 3 = +1 (y-direction) and so on.
at least that's how i interpreted the problem. random walk problems are fun and occur in many branches of science - statistical mechanics, transport problems like heat conduction and diffusion etc.
Last edited by MaxEntropy_Man on Sun Jan 20, 2013 10:04 pm; edited 1 time in total
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
MaxEntropy_Man wrote:atcg wrote:MaxEntropy_Man wrote:and if you roll the die and get 2, you go one unit leftwards, i.e. -x.
Should be -2 leftwards....the word used is "respectively"
atcg: i don't think so. i think the "respectively" is just to keep track of labels. i.e. the jump is always one lattice distance and the matchups are:
1 = +1 (x-direction), 2 = -1 (x-direction), 3 = +1 (y-direction) and so on.
at least that's how i interpreted the problem. random walk problems are fun and occur in many branches of science - statistical mechanics, transport problems like heat conduction and diffusion etc.
Your interpretation is correct. I read the problem again.
Hellsangel- Posts : 14721
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
The answer is that the probability every die roll be unique in the sequence of 6 throws.
6!/(6^6)?
6!/(6^6)?
Hellsangel- Posts : 14721
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Hellsangel wrote:The answer is that the probability every die roll be unique in the sequence of 6 throws.
6!/(6^6)?
on the right track, but there are more possibilities than you have accounted for.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
MaxEntropy_Man wrote:Hellsangel wrote:The answer is that the probability every die roll be unique in the sequence of 6 throws.
6!/(6^6)?
on the right track, but there are more possibilities than you have accounted for.
You are right. You may never get a move in x or y or z direction. Those possibilities need to be added.
Hellsangel- Posts : 14721
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Hellsangel wrote:MaxEntropy_Man wrote:Hellsangel wrote:The answer is that the probability every die roll be unique in the sequence of 6 throws.
6!/(6^6)?
on the right track, but there are more possibilities than you have accounted for.
You are right. You may never get a move in x or y or z direction. Those possibilities need to be added.
no no no. that's not what i meant. every roll of the die results in one of only six possibilities, a unit move in the positive or negative x,y,or z direction. you were right about that. so the denominator of your probability expression is correct. it is your numerator that is not correct. there is more to add there.
edited to add: in retrospect i think that IS what you meant in your original reply. so yeah we are on the same page.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
"0.00167181069958848"
Last edited by atcg on Sun Jan 20, 2013 10:38 pm; edited 1 time in total
Petrichor- Posts : 1725
Join date : 2012-04-10
Re: who's up for a nice meaty probability problem?
bw wrote:155/3888
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
meanwhile, whenever i take an interest in football, new england loses. haven't watched one game this season before tonight. i tune in and they promptly lay an egg. calculate the odds.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Yay for the 49ers! Was on a plane when the game ended, and they gave us a free round of drinks to celebrate.
Interesting problem. I am tempted to solve it using Excel.
Interesting problem. I am tempted to solve it using Excel.
Idéfix- Posts : 8808
Join date : 2012-04-26
Location : Berkeley, CA
Re: who's up for a nice meaty probability problem?
panini press wrote:Yay for the 49ers! Was on a plane when the game ended, and they gave us a free round of drinks to celebrate.
Interesting problem. I am tempted to solve it using Excel.
>>> I was on my way out to check out a movie, but my younger son convinced me to watch the game. Glad I did. The niners didn't play their best game, but had some lucky breaks. Hope they are in better form on the big day.
Kris- Posts : 5461
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Kris wrote:panini press wrote:Yay for the 49ers! Was on a plane when the game ended, and they gave us a free round of drinks to celebrate.
Interesting problem. I am tempted to solve it using Excel.
>>> I was on my way out to check out a movie, but my younger son convinced me to watch the game. Glad I did. The niners didn't play their best game, but had some lucky breaks. Hope they are in better form on the big day.
>>>Forgot to add: kaepernick is definitely a treat to watch
Kris- Posts : 5461
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
I get (60 + 1080 + 720)/(6^6) = 1860/(6^6), which is approximately 4%
Without saying much, just use a symmetry argument and enumerate a small number of cases, and within each case, count the number of distinct sequences of corresponding 6-tuples.
Of course, assuming independent rolls of the standard equi-probable die.
Without saying much, just use a symmetry argument and enumerate a small number of cases, and within each case, count the number of distinct sequences of corresponding 6-tuples.
Of course, assuming independent rolls of the standard equi-probable die.
aanjaneya- Posts : 15
Join date : 2012-09-17
Re: who's up for a nice meaty probability problem?
aanjaneya wrote:I get (60 + 1080 + 720)/(6^6) = 1860/(6^6), which is approximately 4%
Without saying much, just use a symmetry argument and enumerate a small number of cases, and within each case, count the number of distinct sequences of corresponding 6-tuples.
Of course, assuming independent rolls of the standard equi-probable die.
very nice. that's what i got, as did bw. aanjaneya -- are you a professional mathematician?
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
Kris wrote:Kris wrote:panini press wrote:Yay for the 49ers! Was on a plane when the game ended, and they gave us a free round of drinks to celebrate.
Interesting problem. I am tempted to solve it using Excel.
>>> I was on my way out to check out a movie, but my younger son convinced me to watch the game. Glad I did. The niners didn't play their best game, but had some lucky breaks. Hope they are in better form on the big day.
>>>Forgot to add: kaepernick is definitely a treat to watch
overrated - they just have a great front line.
Propagandhi711- Posts : 6941
Join date : 2011-04-29
Re: who's up for a nice meaty probability problem?
MaxEntropy_Man wrote:aanjaneya wrote:I get (60 + 1080 + 720)/(6^6) = 1860/(6^6), which is approximately 4%
Without saying much, just use a symmetry argument and enumerate a small number of cases, and within each case, count the number of distinct sequences of corresponding 6-tuples.
Of course, assuming independent rolls of the standard equi-probable die.
very nice. that's what i got, as did bw. aanjaneya -- are you a professional mathematician?
Thanks for posting this. Glad that our answers agree. Yes, I work in mathematics for a living.
aanjaneya- Posts : 15
Join date : 2012-09-17
Re: who's up for a nice meaty probability problem?
here's another question:
if the same walker takes n steps instead of 6 where n is a very large number, how far is he expected to be from the origin in the end?
if the same walker takes n steps instead of 6 where n is a very large number, how far is he expected to be from the origin in the end?
bw- Posts : 2922
Join date : 2012-11-15
Re: who's up for a nice meaty probability problem?
Intuition says it should be (0,0,0) in expectation, and intuition also says that it should not depend on whether 'n' is large or small.
i.e. one may not need a limiting condition on n \rightarrow \infty.
(Intuition due to the nice symmetry in everything; namely equi-probable faces of the standard die; for every sequence of faces of length n, there exists a neutralizing sequence of length n that also has the same probability; and so on).
In any case, here is one way to work it out.
For 'n' rolls, let the corresponding 6-dimensional vector (n_1, n_2,..., n_6) represent the number of times that the die rolls up as '1', '2', and so on till '6'.
Now (n_1,...n_6) is a random vector that follows a multinomial distribution, with the parameters 'n', and probabilities p_1, p_2,...p_6 = 1/6.
The object of interest is essentially the terminal coordinates in 3-space (i.e. (x,y,z)), which is (n_1-n_2, n_3-n_4, n_5-n_6), after 'n' rolls, where n = \sum_{i=1}^{i=6} n_i
(Due to the +1/-1 moves that specify the stochastic dynamics of this intoxicated particle).
What you are asking is the "expectation" of the above 3-tuple terminal coordinate.
So in other words, you want the expectation of the 3-dimensional random vector (n_1-n_2, n_3-n_4, n_5-n_6), where (n_1, n_2,...n_6) follows a multinomial distribution.
The expectation of each component is zero.
So intuition checks out, I think.
Do post if you agree, or see a flaw.
Thanks.
i.e. one may not need a limiting condition on n \rightarrow \infty.
(Intuition due to the nice symmetry in everything; namely equi-probable faces of the standard die; for every sequence of faces of length n, there exists a neutralizing sequence of length n that also has the same probability; and so on).
In any case, here is one way to work it out.
For 'n' rolls, let the corresponding 6-dimensional vector (n_1, n_2,..., n_6) represent the number of times that the die rolls up as '1', '2', and so on till '6'.
Now (n_1,...n_6) is a random vector that follows a multinomial distribution, with the parameters 'n', and probabilities p_1, p_2,...p_6 = 1/6.
The object of interest is essentially the terminal coordinates in 3-space (i.e. (x,y,z)), which is (n_1-n_2, n_3-n_4, n_5-n_6), after 'n' rolls, where n = \sum_{i=1}^{i=6} n_i
(Due to the +1/-1 moves that specify the stochastic dynamics of this intoxicated particle).
What you are asking is the "expectation" of the above 3-tuple terminal coordinate.
So in other words, you want the expectation of the 3-dimensional random vector (n_1-n_2, n_3-n_4, n_5-n_6), where (n_1, n_2,...n_6) follows a multinomial distribution.
The expectation of each component is zero.
So intuition checks out, I think.
Do post if you agree, or see a flaw.
Thanks.
aanjaneya- Posts : 15
Join date : 2012-09-17
Re: who's up for a nice meaty probability problem?
actually, the distance from the origin is always positive. so i am interested in the expected value of the absolute distance. for example, if a point, constrained to lie on the x axis, is at x=1 with prob 0.5 and at x=-1 with prob 0.5, then its expected distance from the origin is 1(and not 0).
aanjaneya wrote:Intuition says it should be (0,0,0) in expectation, and intuition also says that it should not depend on whether 'n' is large or small.
i.e. one may not need a limiting condition on n \rightarrow \infty.
(Intuition due to the nice symmetry in everything; namely equi-probable faces of the standard die; for every sequence of faces of length n, there exists a neutralizing sequence of length n that also has the same probability; and so on).
In any case, here is one way to work it out.
For 'n' rolls, let the corresponding 6-dimensional vector (n_1, n_2,..., n_6) represent the number of times that the die rolls up as '1', '2', and so on till '6'.
Now (n_1,...n_6) is a random vector that follows a multinomial distribution, with the parameters 'n', and probabilities p_1, p_2,...p_6 = 1/6.
The object of interest is essentially the terminal coordinates in 3-space (i.e. (x,y,z)), which is (n_1-n_2, n_3-n_4, n_5-n_6), after 'n' rolls, where n = \sum_{i=1}^{i=6} n_i
(Due to the +1/-1 moves that specify the stochastic dynamics of this intoxicated particle).
What you are asking is the "expectation" of the above 3-tuple terminal coordinate.
So in other words, you want the expectation of the 3-dimensional random vector (n_1-n_2, n_3-n_4, n_5-n_6), where (n_1, n_2,...n_6) follows a multinomial distribution.
The expectation of each component is zero.
So intuition checks out, I think.
Do post if you agree, or see a flaw.
Thanks.
bw- Posts : 2922
Join date : 2012-11-15
Re: who's up for a nice meaty probability problem?
as long as the jumps are uncorrelated as it is in this case, with the exception of a numerical factor, the displacement should be sqrt(N). i can compute this exactly for the simpler case of 1-d, but need to think harder about 3-d.
MaxEntropy_Man- Posts : 14702
Join date : 2011-04-28
Re: who's up for a nice meaty probability problem?
OK. The terminal coordinate is the 3-tuple, (n_1-n_2, n_3-n_4, n_5-n_6), where (n_1, n_2,...,n_6) follows a multinomial distribution.
So its expectation is indeed (0,0,0).
But bw wants the expectation of the euclidean distance of this point from the origin.
So one needs to compute the Expectation of \sqrt( (n_1-n_2)^2 + (n_3-n_4)^2 + (n_5-n_6)^2 ), with respect to the multinomial distribution of (n_1, n_2,...,n_6).
Using the multinomial distribution properties that Variance(n_i) = np(1-p), and Cov(n_i, n_j) = -np^2,
as well as the following identities, namely,
E(X^2) = Variance(X) + (E(X))^2,
and
E(XY) = Cov(X,Y) + E(X)*E(Y),
using algebra I verify that Expectation of the square of the euclidean distance of the terminal coordinate from the origin simplifies to 6np = n.
So its expectation is indeed (0,0,0).
But bw wants the expectation of the euclidean distance of this point from the origin.
So one needs to compute the Expectation of \sqrt( (n_1-n_2)^2 + (n_3-n_4)^2 + (n_5-n_6)^2 ), with respect to the multinomial distribution of (n_1, n_2,...,n_6).
Using the multinomial distribution properties that Variance(n_i) = np(1-p), and Cov(n_i, n_j) = -np^2,
as well as the following identities, namely,
E(X^2) = Variance(X) + (E(X))^2,
and
E(XY) = Cov(X,Y) + E(X)*E(Y),
using algebra I verify that Expectation of the square of the euclidean distance of the terminal coordinate from the origin simplifies to 6np = n.
aanjaneya- Posts : 15
Join date : 2012-09-17
Re: who's up for a nice meaty probability problem?
one thing to note in addition:
The expectation of the square-distance in this problem is 'n'. This is indeed a very simple closed form expression.
That however does not mean that the expectation of distance is simply \sqrt(n) in closed form.
One needs the variance(distance) characterized in closed form to then further infer what is the closed form of the expectation of distance.
But what you want is indeed characterized by the expectation of \sqrt( (n_1-n_2)^2 + (n_3-n_4)^2 + (n_5-n_6)^2 ), with respect to the multinomial distribution of (n_1, n_2,...,n_6).
So you can indeed write it as a summation, since expectation for a discrete distribution may be expressed as a summation. But you want a closed-form
At first glance, it does not seem all that easy to simplify the above \sqrt to a simple closed-form (as opposed to the square of the above expression, which simplifies to 'n').
Look at the random-walk literature for further pointers. Somebody may have simplified it already for you.
The expectation of the square-distance in this problem is 'n'. This is indeed a very simple closed form expression.
That however does not mean that the expectation of distance is simply \sqrt(n) in closed form.
One needs the variance(distance) characterized in closed form to then further infer what is the closed form of the expectation of distance.
But what you want is indeed characterized by the expectation of \sqrt( (n_1-n_2)^2 + (n_3-n_4)^2 + (n_5-n_6)^2 ), with respect to the multinomial distribution of (n_1, n_2,...,n_6).
So you can indeed write it as a summation, since expectation for a discrete distribution may be expressed as a summation. But you want a closed-form
At first glance, it does not seem all that easy to simplify the above \sqrt to a simple closed-form (as opposed to the square of the above expression, which simplifies to 'n').
Look at the random-walk literature for further pointers. Somebody may have simplified it already for you.
aanjaneya- Posts : 15
Join date : 2012-09-17
Re: who's up for a nice meaty probability problem?
thanks, aanjaneya.
i wonder though if there an easier and more intuitive way of arriving at this answer that i could explain to my 7th grader (right now, he is writing a program in python and will try to plot the results).
i wonder though if there an easier and more intuitive way of arriving at this answer that i could explain to my 7th grader (right now, he is writing a program in python and will try to plot the results).
bw- Posts : 2922
Join date : 2012-11-15
Re: who's up for a nice meaty probability problem?
Awesome. Nothing like a computational experiment to develop and reinforce intuition in most problems.
Python is very very cool.
I will also try to think of a more intuitive explanation, though the above is the most natural one for me. i.e. to think of the quantity of interest as a function of a random variable, whose distribution may be characterized, and then work out its expectation either in closed form (preferable) or numerically.
Python is very very cool.
I will also try to think of a more intuitive explanation, though the above is the most natural one for me. i.e. to think of the quantity of interest as a function of a random variable, whose distribution may be characterized, and then work out its expectation either in closed form (preferable) or numerically.
aanjaneya- Posts : 15
Join date : 2012-09-17
Similar topics
» an extra meaty probability problem
» a little probability problem
» a nice crunchy probability puzzle
» an interesting probability problem
» an interesting probability problem
» a little probability problem
» a nice crunchy probability puzzle
» an interesting probability problem
» an interesting probability problem
Page 1 of 1
Permissions in this forum:
You cannot reply to topics in this forum