[go: up one dir, main page]

Jump to content

Slutsky's theorem: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
 
(41 intermediate revisions by 30 users not shown)
Line 1: Line 1:
{{Short description|Theorem in probability theory}}
{{Refimprove|date=September 2009}}
In [[probability theory]], '''Slutsky’s theorem'''<ref>{{harvnb|Grimmett|2001|loc=Exercise 7.2.5}}</ref> extends some properties of algebraic operations on [[Limit of a sequence|convergent sequences]] of [[real number]]s to sequences of [[random variable]]s.
In [[probability theory]], '''Slutsky's theorem''' extends some properties of algebraic operations on [[Limit of a sequence|convergent sequences]] of [[real number]]s to sequences of [[random variable]]s.<ref>{{cite book |first=Arthur S. |last=Goldberger |author-link=Arthur Goldberger |title=Econometric Theory |location=New York |publisher=Wiley |year=1964 |pages=[https://archive.org/details/econometrictheor0000gold/page/117 117]–120 |url=https://archive.org/details/econometrictheor0000gold |url-access=registration }}</ref>


The theorem was named after [[Eugen Slutsky]].<ref>{{Cite journal
The theorem was named after [[Eugen Slutsky]].<ref>{{harvnb|Slutsky|1925}}</ref> Slutsky’s theorem is also attributed to [[Harald Cramér]].<ref>Slutsky's theorem is also called [[Harald Cramér|Cramér]]’s theorem according to Remark 11.1 (page 249) of Allan Gut. ''A Graduate Course in Probability.'' Springer Verlag. 2005.</ref>
| last = Slutsky | first = E. | author-link = Eugen Slutsky
| year = 1925
| title = Über stochastische Asymptoten und Grenzwerte
| language = de
| journal = Metron
| volume = 5 | issue = 3
| pages = 3–89
| jfm = 51.0380.03
}}</ref> Slutsky's theorem is also attributed to [[Harald Cramér]].<ref>Slutsky's theorem is also called [[Harald Cramér|Cramér]]'s theorem according to Remark 11.1 (page 249) of {{Cite book
| last = Gut | first = Allan
| title = Probability: a graduate course
| publisher = Springer-Verlag
| year = 2005
| isbn = 0-387-22833-0
}}</ref>


==Statement==
==Statement==
Let {''X''<sub>''n''</sub>}, {''Y''<sub>''n''</sub>} be sequences of scalar/vector/matrix [[random element]]s.
Let <math>X_n, Y_n</math> be sequences of scalar/vector/matrix [[random element]]s.
If ''X''<sub>''n''</sub> converges in distribution to a random element ''X'';
If <math>X_n</math> converges in distribution to a random element <math>X</math> and <math>Y_n</math> converges in probability to a constant <math>c</math>, then

and ''Y''<sub>''n''</sub> converges in probability to a constant ''c'', then
* <math>X_n + Y_n \ \xrightarrow{d}\ X + c ;</math>
* <math>X_n + Y_n \ \xrightarrow{d}\ X + c ;</math>
* <math>Y_nX_n \ \xrightarrow{d}\ cX ;</math>
* <math>X_nY_n \ \xrightarrow{d}\ Xc ;</math>
* <math>Y_n^{-1}X_n \ \xrightarrow{d}\ c^{-1}X,</math> &nbsp; provided that ''c'' is invertible,
* <math>X_n/Y_n \ \xrightarrow{d}\ X/c,</math> &nbsp; provided that ''c'' is invertible,


where <math>\xrightarrow{d}</math> denotes [[convergence in distribution]].
where <math>\xrightarrow{d}</math> denotes [[convergence in distribution]].


'''Notes:'''
'''Notes:'''
# The requirement that ''Y<sub>n</sub>'' converges to a constant is important — if it were to converge to a non-degenerate random variable, the theorem would be no longer valid. For example, let <math>X_n \sim {\rm Uniform}(0,1)</math> and <math>Y_n = -X_n</math>. The sum <math>X_n + Y_n = 0</math> for all values of ''n''. Moreover, <math>Y_n \, \xrightarrow{d} \, {\rm Uniform}(-1,0)</math>, but <math>X_n + Y_n</math> does not converge in distribution to <math>X + Y</math>, where <math>X \sim {\rm Uniform}(0,1)</math>, <math>Y \sim {\rm Uniform}(-1,0)</math>, and <math>X</math> and <math>Y</math> are independent.<ref>See {{cite web |first=Donglin |last=Zeng |title=Large Sample Theory of Random Variables (lecture slides) |work=Advanced Probability and Statistical Inference I (BIOS 760) |url=https://www.bios.unc.edu/~dzeng/BIOS760/ChapC_Slide.pdf#page=59 |publisher=University of North Carolina at Chapel Hill |date=Fall 2018 |at=Slide 59 }}</ref>
# In the statement of the theorem, the condition “''Y''<sub>''n''</sub> converges in probability to a constant ''c''” may be replaced with “''Y''<sub>''n''</sub> converges in distribution to a constant ''c''” — these two requirements are equivalent according to [[Convergence of random variables#propB1|this property]].
# The theorem remains valid if we replace all convergences in distribution with convergences in probability.
# The requirement that ''Y<sub>n</sub>'' converges to a constant is important — if it were to converge to a non-degenerate random variable, the theorem would be no longer valid.
# The theorem remains valid if we replace all convergences in distribution with convergences in probability (due to [[Convergence of random variables#propB4|this property]]).


==Proof==
==Proof==
This theorem follows from the fact that if ''X<sub>n</sub>'' converges in distribution to ''X'' and ''Y<sub>n</sub>'' converges in probability to a constant ''c'', then the joint vector (''X<sub>n</sub>, Y<sub>n</sub>'') converges in distribution to (''X, c'') ([[Convergence of random variables#propB3|see here]]).
This theorem follows from the fact that if ''X''<sub>''n''</sub> converges in distribution to ''X'' and ''Y''<sub>''n''</sub> converges in probability to a constant ''c'', then the joint vector (''X''<sub>''n''</sub>, ''Y''<sub>''n''</sub>) converges in distribution to (''X'',&nbsp;''c'') ([[Convergence of random variables#propB3|see here]]).


Next we apply the [[continuous mapping theorem]], recognizing the functions ''g''(''x,y'')=''x+y'', ''g''(''x,y'')=''xy'', and ''g''(''x,y'')=''x''<sup>−1</sup>''y'' as continuous (for the last function to be continuous, ''x'' has to be invertible).
Next we apply the [[continuous mapping theorem]], recognizing the functions ''g''(''x'',''y'')&nbsp;=&nbsp;''x''&nbsp;+&nbsp;''y'', ''g''(''x'',''y'')&nbsp;=&nbsp;''xy'', and ''g''(''x'',''y'')&nbsp;=&nbsp;''x'' ''y''<sup>−1</sup> are continuous (for the last function to be continuous, ''y'' has to be invertible).

==See also==
* [[Convergence of random variables]]


==References==
==References==
{{reflist}}
{{reflist}}

{{refbegin}}
==Further reading==
* {{Citation
* {{cite book |first=George |last=Casella |first2=Roger L. |last2=Berger |title=Statistical Inference |location=Pacific Grove |publisher=Duxbury |year=2001 |pages=240–245 |isbn=0-534-24312-6 }}
* {{Cite book
| last1 = Grimmett | first1 = G.
| last1 = Grimmett | first1 = G.
| last2 = Stirzaker | first2 = D.
| last2 = Stirzaker | first2 = D.
Line 35: Line 54:
| edition = 3rd
| edition = 3rd
}}
}}
* {{cite book |first=Fumio |last=Hayashi |author-link=Fumio Hayashi |title=Econometrics |publisher=Princeton University Press |year=2000 |isbn=0-691-01018-8 |pages=92–93 |url=https://books.google.com/books?id=QyIW8WUIyzcC&pg=PA92 }}
*{{Citation
| last = Gut | first = Allan
| title = Probability: a graduate course
| publisher = Springer-Verlag
| year = 2005
| isbn = 0-387-22833-0
}}
* {{Citation
| last = Slutsky | first = E. | authorlink = Eugen Slutsky
| year = 1925
| title = Über stochastische Asymptoten und Grenzwerte
| language = German
| journal = Metron
| volume = 5 | issue = 3
| pages = 3–89
| jfm = 51.0380.03
}}
{{refend}}


{{DEFAULTSORT:Slutsky's Theorem}}
{{DEFAULTSORT:Slutsky's Theorem}}
[[Category:Asymptotic theory (statistics)]]
[[Category:Probability theorems]]
[[Category:Probability theorems]]
[[Category:Statistical theorems]]
[[Category:Theorems in statistics]]

Latest revision as of 16:44, 26 November 2023

In probability theory, Slutsky's theorem extends some properties of algebraic operations on convergent sequences of real numbers to sequences of random variables.[1]

The theorem was named after Eugen Slutsky.[2] Slutsky's theorem is also attributed to Harald Cramér.[3]

Statement

[edit]

Let be sequences of scalar/vector/matrix random elements. If converges in distribution to a random element and converges in probability to a constant , then

  •   provided that c is invertible,

where denotes convergence in distribution.

Notes:

  1. The requirement that Yn converges to a constant is important — if it were to converge to a non-degenerate random variable, the theorem would be no longer valid. For example, let and . The sum for all values of n. Moreover, , but does not converge in distribution to , where , , and and are independent.[4]
  2. The theorem remains valid if we replace all convergences in distribution with convergences in probability.

Proof

[edit]

This theorem follows from the fact that if Xn converges in distribution to X and Yn converges in probability to a constant c, then the joint vector (Xn, Yn) converges in distribution to (Xc) (see here).

Next we apply the continuous mapping theorem, recognizing the functions g(x,y) = x + y, g(x,y) = xy, and g(x,y) = x y−1 are continuous (for the last function to be continuous, y has to be invertible).

See also

[edit]

References

[edit]
  1. ^ Goldberger, Arthur S. (1964). Econometric Theory. New York: Wiley. pp. 117–120.
  2. ^ Slutsky, E. (1925). "Über stochastische Asymptoten und Grenzwerte". Metron (in German). 5 (3): 3–89. JFM 51.0380.03.
  3. ^ Slutsky's theorem is also called Cramér's theorem according to Remark 11.1 (page 249) of Gut, Allan (2005). Probability: a graduate course. Springer-Verlag. ISBN 0-387-22833-0.
  4. ^ See Zeng, Donglin (Fall 2018). "Large Sample Theory of Random Variables (lecture slides)" (PDF). Advanced Probability and Statistical Inference I (BIOS 760). University of North Carolina at Chapel Hill. Slide 59.

Further reading

[edit]
  • Casella, George; Berger, Roger L. (2001). Statistical Inference. Pacific Grove: Duxbury. pp. 240–245. ISBN 0-534-24312-6.
  • Grimmett, G.; Stirzaker, D. (2001). Probability and Random Processes (3rd ed.). Oxford.
  • Hayashi, Fumio (2000). Econometrics. Princeton University Press. pp. 92–93. ISBN 0-691-01018-8.