**Basic Probability Theory**

Intuitively, a random variable (r.v.) *X* is a function mapping its sample space to the set of real numbers.

**Exercise:** Read about sample spaces.

**(a)** A fair six-sided dice is tossed. What is a "natural" sample space? What is a "natural" definition for the corresponding random variable.

**(b)** Two fair six-sided dies are tossed. Define the sample space to be the set of possible values for the sum of the values on the two dies. What is the "natural" definition for the corresponding random variable?

**Definition:** The cumulative distribution function (c.d.f.) of a real-valued random variable represents the probability that the random variable *X* takes on a value less than or equal to *x.* The c.d.f. can be defined for both discrete and continuous real-valued random variables.

**Definition:** The probability mass function (p.m.f.) of a discrete random variable represents the probability that the random variable *X* takes on a value less than or equal to *x.*

**Definition:** The probability density function (p.d.f.) of a discrete or continuous (or mixed!) random variable is an analogue of the p.m.f. -- the integral of the p.d.f. over any interval represents the probability that the random variable *X* takes on a value in the interval*, i.e.,* *.*

**Exercise:** Read about joint distributions and conditional distributions of sets of random variables. Construct two non-trivial examples of each.

**Exercise:** Construct c.d.f.s and p.d.f.s for any two examples of continous random variables, and c.d.f.s, p.m.f.s and p.d.f.s for any two examples of discrete random variables.

**Definition:** A discrete-time stochastic process (the only kind of stochastic processes we'll consider in this class) is just a sequence of random variables.

**Exercise:** Read about (strictly) stationary stochastic processes (intuitively, in a stationary process the joint distributions of any subsets of random variables are identical, *i.e.*, stationary). Construct two examples of stationary stochastic processes, and two examples of non-stationary stochastic processes.

**Exercise:** Read about ergodic stochastic processes (intuitively, in an ergodic process a sequence of *n* observations of a random variable behaves in exactly the same manner as *n* copies of that random variable). Construct two examples of ergodic processes, and two examples of non-ergodic processes.

**Exercise:** Read about independent and conditionally independent random variables (intuitively, two r.v.s *X* and *Y* are independent if the probability of events in one's sample space is independent of any event in the other's sample space; two r.v.s *X* and *Y* are conditionally independent given a third r.v. *Z* if the probability of events in *X*'s sample space is independent of any event in Y's sample space (and vice versa), given that the value of *Z* is fixed and known). For three (or more) random variables, what are the appropriate definitions? *(Hint -- the number of conditions that have to be satisfied grows exponentially with the number of random variables)*. Give examples of a sequence of *n* random variables such that any *n-1* are independent, but all *n* together are not. Can you generalize this to an example where, for any *k < n*, any set of *k* r.v.s are independent, but any set of *k+1* r.v.s are not?

**Exercise:** Read about Markov chains (intuitively, a Markov chain is a sequence of random variables such at any instant, the past and the future are conditionally independent given the present). Give two examples of Markov chains.

## Comments (0)

You don't have permission to comment on this page.