# Difference between revisions of "CDS 110b: Random Processes"

 WARNING: This page is for a previous year. See current course homepage to find most recent page available.

This lecture presents an introduction to random processes.

## Lecture Outline

1. Quick review of random variables
2. Random Processes
• Definition
• Properties
• Example

## References and Further Reading

• Hoel, Port and Stone, Introduction to Probability Theory - this is a good reference for basic definitions of random variables
• Apostol II, Chapter 14 - another reference for basic definitions in probability and random variables

## Frequently Asked Questions

Q: Can you explain the jump from pdfs to correlations in more detail?

The probability density function (pdf), ${\displaystyle p(x;t)}$ tells us how the value of a random process is distributed at a particular time:

${\displaystyle P(a\leq x(t)\leq b)=\int _{a}^{b}p(x;t)dx.}$

You can interpret this by thinking of ${\displaystyle x(t)}$ as a separate random variable for each time ${\displaystyle t}$

The correlation for a random process tells us how the value of a random process at one time, ${\displaystyle t_{1}}$ is related to the value at a different time ${\displaystyle t_{2}}$. This relationship is probabalistic, so it is also described in terms of a distribution. In particular, we use the joint probability density function, ${\displaystyle p(x_{1},x_{2};t_{1},t_{2})}$ to characterize this:

${\displaystyle P(a_{1}\leq x_{1}(t_{1})\leq b_{1},a_{2}\leq x_{2}(t_{2})\leq b_{2})=\int _{a_{1}}^{b_{1}}\int _{a_{2}}^{b_{2}}p(x_{1},x_{2};t_{1},t_{2})dx_{1}dx_{2}}$

Given any random process, ${\displaystyle p(x_{1},x_{2};t_{1},t_{2})}$ descibes (as a density) how the value of the random variable at time ${\displaystyle t_{1}}$ is related (or "correlated") with the value at time ${\displaystyle t_{2}}$. We can thus describe a random process according to its joint probability density function.

In practice, we don't usually describe random processes in terms of their pdfs and joint pdfs. It is usually easier to describe them in terms of their statistics (mean, variance, etc). In particular, we almost never describe the correlation in terms of joint pdfs, but instead use the correlation function:

${\displaystyle \rho (t,\tau )=E\{x(t)x(\tau )\}=\int _{-\infty }^{\infty }\int _{-\infty }^{\infty }x_{1}x_{2}p(x_{1},x_{2};t,\tau )dx_{1}dx_{2}}$

The utility of this particular function is seen primarily through its application: if we know the correlation for one random process and we "filter" that random process through a linear system, we can compute the correlation for the corresponding output process (we'll see this in the next lecture).