5.1: Joint Distributions of Discrete Random Variables (2024)

  1. Last updated
  2. Save as PDF
  • Page ID
    3272
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vectorC}[1]{\textbf{#1}}\)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}}\)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}}\)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    In this chapter we consider two or more random variables defined on the same sample space and discuss how to model the probability distribution of the random variables jointly. We will begin with the discrete case by looking at the joint probability mass function for two discrete random variables. In the following section, we will considercontinuous random variables.

    Definition \(\PageIndex{1}\)

    If discrete random variables \(X\) and \(Y\) are defined on the same sample space \(S\), then their joint probability mass function(joint pmf) is given by
    $$p(x,y) = P(X=x\ \ \text{and}\ \ Y=y),\notag$$
    where \((x,y)\) is a pair of possible values for the pair of random variables \((X,Y)\), and \(p(x,y)\) satisfies the following conditions:

    • \(0 \leq p(x,y) \leq 1\)
    • \(\displaystyle{\mathop{\sum\sum}_{(x,y)}p(x,y) = 1}\)
    • \(\displaystyle{P\left((X,Y)\in A\right)) = \mathop{\sum\sum}_{(x,y)\in A} p(x,y)}\)

    Note that conditions #1 and #2 in Definition 5.1.1 are required for \(p(x,y)\) to be a valid joint pmf, while the third condition tells us how to use the joint pmf to find probabilities for the pair of random variables \((X,Y)\).

    In the discrete case, we can obtain the joint cumulative distribution function (joint cdf)of \(X\) and \(Y\) by summing the joint pmf:
    $$F(x,y) = P(X\leq x\ \text{and}\ Y\leq y) = \sum_{x_i \leq x} \sum_{y_j \leq y} p(x_i, y_j),\notag$$
    where \(x_i\) denotes possible values of \(X\) and \(y_j\) denotes possible values of \(Y\). From the joint pmf, we can also obtain the individual probability distributions of \(X\) and \(Y\) separately as shown in the next definition.

    Definition \(\PageIndex{2}\)

    Suppose that discrete random variables \(X\) and \(Y\) have joint pmf\(p(x,y)\). Let \(x_1, x_2, \ldots, x_i, \ldots\) denote the possible values of \(X\), and let \(y_1, y_2, \ldots, y_j, \ldots\) denote the possible values of \(Y\). The marginal probability massfunctions(marginal pmf's)of \(X\) and \(Y\) are respectively given by the following:
    \begin{align*}
    p_X(x) &= \sum_j p(x, y_j) \quad(\text{fix a value of}\ X\ \text{and sum over possible values of}\ Y) \\
    p_Y(y) &= \sum_i p(x_i, y) \quad(\text{fix a value of}\ Y\ \text{and sum over possible values of}\ X)
    \end{align*}

    Example \(\PageIndex{1}\)

    Consider again the probability experiment of Example 3.3.2, where we toss a fair coin three times and record the sequence of heads \((h)\) and tails \((t)\). Again, we let random variable \(X\) denote the number of heads obtained. We also let random variable \(Y\) denote the winnings earned in a single play of a game with the following rules, based on the outcomes of the probability experiment (this is the same as Example 3.6.2):

    • player wins $1 if first \(h\) occurs on the first toss
    • player wins $2 if first \(h\) occurs on the second toss
    • player wins $3 if first \(h\) occurs on the third toss
    • player loses $1 if no \(h\) occur

    Note that the possible values of \(X\) are \(x=0,1,2,3\), and the possible values of \(Y\) are \(y=-1,1,2,3\). We represent the joint pmfusing a table:

    Table 1: joint pmf of \(X\) and \(Y\)

    \(p(x,y)\) \(X\)
    \(Y\) 0 1 2 3
    -1 1/8 0 0 0
    1 0 1/8 2/8 1/8
    2 0 1/8 1/8 0
    3 0 1/8 0 0

    The values in Table 1 give the values of \(p(x,y)\). For example, consider \(p(0,-1)\):
    $$p(0,-1) = P(X=0\ \text{and}\ Y=-1) = P(ttt) = \frac{1}{8}.\notag$$
    Since the outcomes are equally likely, the values of \(p(x,y)\) are found by counting the number of outcomes in the sample space \(S\)that result in the specified values of the random variables, and then dividing by \(8\), the total number of outcomes in \(S\). The sample space is given below, color coded to help explain the values of \(p(x,y)\):
    $$S= \{{\color{green}ttt}, {\color{orange}htt}, {\color{orange}tht}, {\color{orange}tth}, {\color{blue}hht}, {\color{blue}hth}, {\color{blue}thh}, {\color{purple} hhh}\}\notag$$

    Given the joint pmf, we can now find the marginal pmf's. Note that the marginal pmffor \(X\) is found by computing sums of the columns in Table 1, and the marginal pmffor \(Y\) corresponds to the row sums. (Note that we found the pmffor \(X\) in Example 3.3.2as well, it is a binomial random variable. We also found the pmf for \(Y\) in Example 3.6.2.)

    Table 2: marginal pmf's for \(X\) and \(Y\)

    \(x\) \(p_X(x)\) \(y\) \(p_Y(y)\)
    0 1/8 -1 1/8
    1 3/8 1 1/2
    2 3/8 2 1/4
    3 1/8 3 1/8

    Finally, we can find the joint cdf for \(X\) and \(Y\) by summing over values of the joint frequency function. For example, consider \(F(1,1)\):
    $$F(1,1) = P(X\leq1\ \text{and}\ Y\leq1) = \sum_{x\leq1}\sum_{y\leq1} p(x,y) =p(0,-1) + p(0,1) + p(-1,1) + p(1,1) = \frac{1}{4}\notag$$
    Again, we can represent the joint cdf using a table:

    Table 3: joint cdf of \(X\) and \(Y\)

    \(F(x,y)\) \(X\)
    \(Y\) 0 1 2 3
    -1 1/8 1/8 1/8 1/8
    1 1/8 1/4 1/2 5/8
    2 1/8 3/8 3/4 7/8
    3 1/8 1/2 7/8 1

    Expectations of Functions of Jointly Distributed Discrete Random Variables

    We now look at taking the expectation of jointly distributed discrete random variables. Because expected values are defined for a single quantity, we will actually define the expected value of a combination of the pair of random variables, i.e., we look at the expected value of a function applied to \((X,Y)\).

    Theorem \(\PageIndex{1}\)

    Suppose that \(X\) and \(Y\) are jointly distributed discrete random variables with joint pmf \(p(x,y)\).

    If \(g(X,Y)\) is a function of these two random variables, then its expected value is given by the following:
    $$\text{E}[g(X,Y)] = \mathop{\sum\sum}_{(x,y)}g(x,y)p(x,y).\notag$$

    Example \(\PageIndex{2}\)

    Consider again the discrete random variables we defined in Example 5.1.1 with joint pmf given in Table 1. We will find the expected value of three different functions applied to \((X,Y)\).

    1. First, we define \(g(x,y) = xy\), and compute the expected value of \(XY\):
      \begin{align*}
      \text{E}[XY] = \mathop{\sum\sum}_{(x,y)}xy\cdot p(x,y) &= (0)(-1)\left(\frac{1}{8}\right) \\
      &\ + (1)(1)\left(\frac{1}{8}\right) + (2)(1)\left(\frac{2}{8}\right) + (3)(1)\left(\frac{1}{8}\right) \\
      &\ + (1)(2)\left(\frac{1}{8}\right) + (2)(2)\left(\frac{1}{8}\right) \\
      &\ + (1)(3)\left(\frac{1}{8}\right) \\
      &= \frac{17}{8} = 2.125
      \end{align*}
    2. Next, we define \(g(x) = x\), and compute the expected value of \(X\):
      \begin{align*}
      \text{E}[X] = \mathop{\sum\sum}_{(x,y)}x\cdot p(x,y) &= (0)\left(\frac{1}{8}\right) \\
      &\ + (1)\left(\frac{1}{8}\right) + (2)\left(\frac{2}{8}\right) + (3)\left(\frac{1}{8}\right) \\
      &\ + (1)\left(\frac{1}{8}\right) + (2)\left(\frac{1}{8}\right) \\
      &\ + (1)\left(\frac{1}{8}\right)\\
      &= \frac{12}{8} = 1.5
      \end{align*}
      Recall that \(X\sim\text{binomial}(n = 3, p = 0.5)\), and that the expected value of a binomial random variable is given by \(np\). Thus, we can verify the expected value of \(X\) that we calculated above using Theorem 5.1.1 using this fact for binomial distributions: \(\text{E}[X] = np = 3(0.5) = 1.5\).
    3. Lastly, we define \(g(x,y) = y\), and calculate the expected value of \(Y\):
      \begin{align*}
      \text{E}[Y] = \mathop{\sum\sum}_{(x,y)}y\cdot p(x,y) &= (-1)\left(\frac{1}{8}\right) \\
      &\ + (1)\left(\frac{1}{8}\right) + (1)\left(\frac{2}{8}\right) + (1)\left(\frac{1}{8}\right) \\
      &\ + (2)\left(\frac{1}{8}\right) + (2)\left(\frac{1}{8}\right) \\
      &\ + (3)\left(\frac{1}{8}\right) \\
      &= \frac{10}{8} = 1.25
      \end{align*}
      Again, we can verify this result by reviewing the calculations done in Example 3.6.2.

    Independent Random Variables

    In some cases, the probability distribution of one random variable will not be affected by the distribution of another random variable defined on the same sample space. In those cases, the joint distribution functions have a very simple form, and we refer to the random variables as independent.

    Definition \(\PageIndex{3}\)

    Discrete random variables \(X_1, X_2, \ldots, X_n\) are independent if the joint pmf factors into a product of the marginal pmf's:
    $$p(x_1, x_2, \ldots, x_n) = p_{X_1}(x_1)\cdot p_{X_2}(x_2) \cdots p_{X_n}(x_n).\label{indeprvs}$$
    It is equivalent to check that this condition holds for the cumulative distribution functions.

    Recall the definition of independentevents(Definition 2.3.2): \(A\) and \(B\) are independent events if \(P(A\cap B) = P(A)\ P(B)\). This is the basis for the definition of independent random variables because we can write the pmf's in Equation \ref{indeprvs} in terms of events as follows:
    $$p(x,y) = P(X=x\ \text{and}\ Y=y) = P(\{X=x\}\cap\{Y=y\}) = P(X=x) P(Y=y) = p_X(x) p_Y(y)\notag$$
    In the above, we use the idea that if \(X\) and \(Y\) are independent, then the event that \(X\) takes on a given value \(x\) is independent of the event that \(Y\) takes the value \(y\).

    Example \(\PageIndex{3}\)

    Consider yet againthe discrete random variables defined in Example 5.1.1. According to the definition,\(X\) and \(Y\) are independent if
    $$p(x,y) = p_X(x)\cdot p_Y(y),\notag$$
    for all pairs \((x,y)\). Recall that the joint pmffor \((X,Y)\) is given in Table 1 and that themarginal pmf's for \(X\) and \(Y\) are given in Table 2.Note that, for \((x,y) = (0,-1)\), we have the following
    $$p(0,-1) = \frac{1}{8},\ \ p_X(0) = \frac{1}{8},\ \ p_Y(-1) = \frac{1}{8} \quad\Rightarrow\quad p(0,-1) \neq p_X(0)\cdot p_Y(-1).\notag$$
    Thus, \(X\) and \(Y\) are not independent, or in other words, \(X\) and \(Y\) are dependent. This should make sense given the definition of \(X\) and \(Y\). The winnings earned depend on the number of heads obtained. So the probabilities assigned to the values of \(Y\) will be affected by the values of \(X\).

    We also have the following very useful theorem about the expected value of a product of independent random variables, which is simply given by the product of the expected values for the individual random variables.

    Theorem \(\PageIndex{2}\)

    If \(X\) and \(Y\) are independent random variables, then \(\text{E}[XY] = \text{E}[X]\ \text{E}[Y]\).

    Proof

    Assume \(X\) and \(Y\) are independent random variables. If we let \(p(x,y)\) denote the joint pmf of \((X, Y)\), then, by Definition 5.1.3, \(p(x,y) = p_X(x)p_Y(y)\), for all pairs \((x,y)\). Using this fact and Theorem 5.1.1, we have
    \begin{align*}
    \text{E}[XY] &= \mathop{\sum\sum}_{(x,y)}xy\cdot p(x,y) = \mathop{\sum\sum}_{(x,y)}xy\cdot p_X(x)p_Y(y)\\
    &= \sum_x\sum_y xyp_(x)P_Y(y) = \sum_x xp_X(x) \left(\sum_y p_Y(y)\right) = \sum_x xp_X(x)\text{E}[Y] \\
    &= \text{E}[Y]\sum_x xp_X(x) = \text{E}[Y]\ \text{E}[X].
    \end{align*}

    Theorem 5.1.2 can be used to show that two random variables arenotindependent:if \(\text{E}[XY] \neq \text{E}[X]\ \text{E}[Y]\), then \(X\) and \(Y\)cannotbe independent. However, beware using Theorem 5.1.2 to show that random variables are independent. Note that Theorem 5.1.2assumesthat \(X\) and \(Y\) are independent and then the property about the expected value follows. The other direction does not hold. In other words, if\(\text{E}[XY] =\text{E}[X]\ \text{E}[Y]\), then \(X\) and \(Y\)may or may notbe independent.

    5.1: Joint Distributions of Discrete Random Variables (2024)
    Top Articles
    Aptilink IQ Test: A Comprehensive Review IQ Limit
    Darwin Núñez Net Worth
    11 beste sites voor Word-labelsjablonen (2024) [GRATIS]
    Wordscapes Level 6030
    Uihc Family Medicine
    Chelsea player who left on a free is now worth more than Palmer & Caicedo
    Botanist Workbench Rs3
    1TamilMV.prof: Exploring the latest in Tamil entertainment - Ninewall
    Best Restaurants In Seaside Heights Nj
    Housing Intranet Unt
    Keurig Refillable Pods Walmart
    Persona 4 Golden Taotie Fusion Calculator
    Pwc Transparency Report
    OSRS Dryness Calculator - GEGCalculators
    Colts seventh rotation of thin secondary raises concerns on roster evaluation
    Lima Funeral Home Bristol Ri Obituaries
    Vistatech Quadcopter Drone With Camera Reviews
    The Grand Canyon main water line has broken dozens of times. Why is it getting a major fix only now?
    Touchless Car Wash Schaumburg
    Rs3 Eldritch Crossbow
    Slim Thug’s Wealth and Wellness: A Journey Beyond Music
    Jayme's Upscale Resale Abilene Photos
    Netspend Ssi Deposit Dates For 2022 November
    Umn Biology
    Federal Express Drop Off Center Near Me
    LG UN90 65" 4K Smart UHD TV - 65UN9000AUJ | LG CA
    Craigslist/Phx
    Nacogdoches, Texas: Step Back in Time in Texas' Oldest Town
    Gabrielle Enright Weight Loss
    Weekly Math Review Q4 3
    Tamil Play.com
    Strange World Showtimes Near Atlas Cinemas Great Lakes Stadium 16
    Atlantic Broadband Email Login Pronto
    Craigslist Greencastle
    Santa Cruz California Craigslist
    How Much Is Mink V3
    دانلود سریال خاندان اژدها دیجی موویز
    Kerry Cassidy Portal
    303-615-0055
    Tunica Inmate Roster Release
    Ds Cuts Saugus
    Courses In Touch
    Cch Staffnet
    Canada Life Insurance Comparison Ivari Vs Sun Life
    Joblink Maine
    855-539-4712
    Kenwood M-918DAB-H Heim-Audio-Mikrosystem DAB, DAB+, FM 10 W Bluetooth von expert Technomarkt
    Union Supply Direct Wisconsin
    Rovert Wrestling
    Motorcycle For Sale In Deep East Texas By Owner
    Is My Sister Toxic Quiz
    Greg Steube Height
    Latest Posts
    Article information

    Author: Gov. Deandrea McKenzie

    Last Updated:

    Views: 5797

    Rating: 4.6 / 5 (46 voted)

    Reviews: 93% of readers found this page helpful

    Author information

    Name: Gov. Deandrea McKenzie

    Birthday: 2001-01-17

    Address: Suite 769 2454 Marsha Coves, Debbieton, MS 95002

    Phone: +813077629322

    Job: Real-Estate Executive

    Hobby: Archery, Metal detecting, Kitesurfing, Genealogy, Kitesurfing, Calligraphy, Roller skating

    Introduction: My name is Gov. Deandrea McKenzie, I am a spotless, clean, glamorous, sparkling, adventurous, nice, brainy person who loves writing and wants to share my knowledge and understanding with you.