Talk:Stochastic convergence: Difference between revisions
imported>Robert Tito |
imported>Greg Woodhouse (Almost sure convergence - response) |
||
Line 21: | Line 21: | ||
The definition in the text doesn't make sense. In general, something is true almost surely (a.s.) if it is true with a probability of 1. It is almost surely true that a randomly chosen number is not 4. [[User:Greg Woodhouse|Greg Woodhouse]] 12:08, 28 June 2007 (CDT) | The definition in the text doesn't make sense. In general, something is true almost surely (a.s.) if it is true with a probability of 1. It is almost surely true that a randomly chosen number is not 4. [[User:Greg Woodhouse|Greg Woodhouse]] 12:08, 28 June 2007 (CDT) | ||
:Greg, some would argue if the probability has a chance of being not true ≤0.01‰ the convergenge to a value is true, and the tail can be forgotten. Compare to series. [[User:Robert Tito|Robert Tito]] | <span style="background:grey"> <font color="yellow"><b>[[User talk:Robert Tito|Talk]]</b></font> </span> 12:44, 28 June 2007 (CDT) | :Greg, some would argue if the probability has a chance of being not true ≤0.01‰ the convergenge to a value is true, and the tail can be forgotten. Compare to series. [[User:Robert Tito|Robert Tito]] | <span style="background:grey"> <font color="yellow"><b>[[User talk:Robert Tito|Talk]]</b></font> </span> 12:44, 28 June 2007 (CDT) | ||
I understand. It's a technical concept from measure theory. If two functions (say ''f'' and ''g'') are equal except on a set of measure 0, we say ''f = g'' almost everywhere. This is important because their (Lebesgue) integrals over a given set will always be equal and it is convenient to ''identify'' such functions because then if we define | |||
:<math>(f,g) = \int_A f\,\bar{g}\,dx</math> | |||
defines a metric on <math>\scriptstyle L^2(A)</math>, giving it the structure of a metric space. If we didn't identify functions that wee equal almost everywhere (i.e., treat them as the same function), the purported metric would not be positive definite. | |||
On a foundational level, probability theory is essentially measure theory, (and distributions are just measurable functions that, when integrated ovderf a set ''A'' give the probability of ''A''). It is just a convention that probability theorists use the phrase "almost surely" instead of "almost everywhere", the meanings are the same. Of course, I'm almost sure :) you already know this! [[User:Greg Woodhouse|Greg Woodhouse]] 13:28, 28 June 2007 (CDT) |
Revision as of 12:28, 28 June 2007
Workgroup category or categories | Mathematics Workgroup, Physics Workgroup, Chemistry Workgroup [Categories OK] |
Article status | Stub: no more than a few sentences |
Underlinked article? | Not specified |
Basic cleanup done? | Yes |
Checklist last edited by | Ragnar Schroder 11:13, 28 June 2007 (CDT) |
To learn how to fill out this checklist, please see CZ:The Article Checklist.
Work in progress
This page is a work in progress, I'm struggling with the latex and some other stuff, there may be actual errors now.
Ragnar Schroder 12:04, 28 June 2007 (CDT)
Almost sure convergence
The definition in the text doesn't make sense. In general, something is true almost surely (a.s.) if it is true with a probability of 1. It is almost surely true that a randomly chosen number is not 4. Greg Woodhouse 12:08, 28 June 2007 (CDT)
- Greg, some would argue if the probability has a chance of being not true ≤0.01‰ the convergenge to a value is true, and the tail can be forgotten. Compare to series. Robert Tito | Talk 12:44, 28 June 2007 (CDT)
I understand. It's a technical concept from measure theory. If two functions (say f and g) are equal except on a set of measure 0, we say f = g almost everywhere. This is important because their (Lebesgue) integrals over a given set will always be equal and it is convenient to identify such functions because then if we define
defines a metric on , giving it the structure of a metric space. If we didn't identify functions that wee equal almost everywhere (i.e., treat them as the same function), the purported metric would not be positive definite.
On a foundational level, probability theory is essentially measure theory, (and distributions are just measurable functions that, when integrated ovderf a set A give the probability of A). It is just a convention that probability theorists use the phrase "almost surely" instead of "almost everywhere", the meanings are the same. Of course, I'm almost sure :) you already know this! Greg Woodhouse 13:28, 28 June 2007 (CDT)
- Mathematics Category Check
- General Category Check
- Physics Category Check
- Chemistry Category Check
- Advanced Articles
- Nonstub Articles
- Internal Articles
- Mathematics Advanced Articles
- Mathematics Nonstub Articles
- Mathematics Internal Articles
- Physics Advanced Articles
- Physics Nonstub Articles
- Physics Internal Articles
- Chemistry Advanced Articles
- Chemistry Nonstub Articles
- Chemistry Internal Articles
- Developed Articles
- Mathematics Developed Articles
- Physics Developed Articles
- Chemistry Developed Articles
- Developing Articles
- Mathematics Developing Articles
- Physics Developing Articles
- Chemistry Developing Articles
- Stub Articles
- Mathematics Stub Articles
- Physics Stub Articles
- Chemistry Stub Articles
- External Articles
- Mathematics External Articles
- Physics External Articles
- Chemistry External Articles
- Mathematics Underlinked Articles
- Underlinked Articles
- Physics Underlinked Articles
- Chemistry Underlinked Articles
- Mathematics Cleanup
- General Cleanup
- Physics Cleanup
- Chemistry Cleanup