Any signal that can be uniquely described by an explicit mathematical expression, a well-defined rule or a table look-up, is called a deterministic signal. The value of such signals is known or determined precisely at every instant of time. In many practical applications, however, some signals are generated in a random fashion and cannot be explicitly described prior to their occurrence. Such signals are referred to as nondeterministic or random signals. An example of random signal is noise, an ever-present undesirable signal that contaminates the message signal during its passage through a communication link and makes the information erroneous. The composite signal, the desired signal plus interfering noise components at the input of the receiver, is again random in nature. Although there is an uncertainty about the exact nature of the random signals, probabilistic methods can be adopted to present their general behaviour. This chapter aims at presenting a short course on theory of probability, random processes and noise, essential for extracting information-bearing signals from noisy background.
Probability is a measure of certainty. The theory of probability originated from the analysis of certain games of chance, such as roulette and cards. Soon it became a powerful mathematical tool almost in all branches of science and engineering. The following terms are useful in developing the concept of probability:
Experiment—Any process of observation is known as an experiment.
Outcome—It is the result of an experiment.
Random Experiment—An experiment is called a random experiment if its outcome is not unique and therefore cannot be predicted with certainty. Typical examples of a random experiment are tossing a coin, rolling a die or selecting a message from several messages for transmission.
Sample Space, Sample Point and Event—The set S of all possible outcomes of a given random experiment is called the sample space, universal set, or certain event. An element (outcome) in S is known as a sample point or elementary event. An event A is a subset of a sample space (A ⊂ S).
In any random experiment, we are not certain as to whether a particular event will occur or not. As a measure of the probability, we usually assign a number between 0 and 1. If we are sure that the event will occur, the probability of that event is taken as 1 (100%). When the event will never occur, the probability of that event is 0 (0%). If the probability is 9/10, there is a 90% chance that it will occur. In other words, there is a 10% (1/10) probability that the event will not occur.
The probability of an event can be defined by the following two approaches:
However, both the approaches have serious drawbacks, as the terms ‘equally likely’ and ‘large number’ are vague. Due to these drawbacks, an axiomatic approach was developed, which is described in Sec. 1.2.2.
We can combine events to form new events using various set operations as follows:
Example 1.1: Throw a die and observe the number of dots appearing on the top face. (a) Construct its sample space. (b) If A be the event that an odd number occurs, B that an even number occurs, and C that a prime number occurs, write down their respective subsets. (c) Find the event that an even or a prime number occurs. (d) List the outcomes for the event that an even prime number occurs. (e) Find the event that a prime number does not occur. (f) Find the event that seven dots appear on the top face. (g) Find the event that even and odd numbers occur simultaneously.
Solution:
It has six sample points.
B = {even number} = {2,4,6}
C = {prime number} = {2,3,5}.

Let P(A) denote the probability of event A of some sample space S. It must satisfy the following axioms:

Some important theorems on probability are as follows:

Example 1.2: Determine the probability for the event that the sum 8 appears in a single throw of a pair of fair dice.
Solution: The sum 8 appears in the following cases:
(2,6), (3,5), (4,4), (5,3), (6,2), i.e., 5 cases
Total number of outcomes is 6 × 6 = 36
Thus, 
Example 1.3: A ball is drawn at random from a box containing 5 red balls, 4 green balls, and 3 blue balls. Determine the probability that it is (a) green, (b) not green, and (c) red or blue.
Solution:
Let R, G, and B denote the events of drawing a red ball, green ball, and blue ball, respectively.
Then 
Method 2
The sample space consists of 12 sample points. If we assign equal probabilities 1/12 to each sample point, we again see that P(G) = 4/12 = 1/3, since there are 4 sample points corresponding to green ball.

Method 2
Method 3
Since events R and G are mutually exclusive, we have
Example 1.4: Two cards are drawn at random from a pack of 52 cards. Find the probability that (a) both are hearts and (b) one is heart and one is spade.
Solution: There are
ways that 2 cards can be drawn from 52 cards.
ways to draw 2 hearts from 13 hearts.Thus, the probability that both cards are hearts = 
Thus, the probability that one is heart and one is spade 
Example 1.5: A telegraph source emits two symbols, dash and dot. It was observed that the dash were thrice as likely to occur as dots. Find the probabilities of the dashes and dots occurring.
Solution: We have, P(dash) = 3P(dot).
Now the sum of the probabilities must be 1.
Example 1.6: A digital message is 1000 symbols long. An average of 0.1% symbols received is erroneous. (a) What is the probability of correct reception? (b) How many symbols are correctly received?
Solution:
Thus, the probability of correct reception = P(Ā) = 1 − P(A) = 1 − 0.001 = 0.999
The conditional probability of an event A given that B has happened (P(A/B)) is defined as follows:
where P(A ∩ B) is the joint probability of A and B.
Similarly,
Using Eqs. (1.9) and (1.10) we can write
or
This is known as Bayes rule.
Example 1.7: Consider the following table:
(a) If a person is male, what is the probability that he is unemployed? (b) If a person is female, what is the probability that she is employed? (c) If a person is employed, what is the probability that he is male?
Solution:



Two events A and B are said to be (statistically) independent if
or
i.e., the occurrence (or non-occurrence) of event A has no influence on the occurrence (or non-occurrence) of B. Otherwise they are said to be dependent.
Combining Eqs. (1.11) and (1.13), we have
Example 1.8: Determine the probability for the event that at least one head appears in three tosses of a fair coin.
Example 1.9: An experiment consists of observation of five successive pulse positions on a communication link. The pulses can be positive, negative or zero. Also consider that the individual experiments that determine the kind of pulse at each possible position are independent. If the probabilities of occurring a positive pulse and a negative pulse at the ith position (xi) are p = 0.5 and q = 0.2, respectively, then find (a) the probability that all pulses are negative and (b) the probability that the first two pulses are positive, the next two are zero and the last is negative.
Solution:

Hence, the probability that the first two pulses are positive, the next two are zero and the last is negative is given by
Let A1, A2, …, An be mutually exclusive (Ai ∩ Aj = Ø, for i ≠ j) and exhaustive
Now B is any event in S.
Then
This is known as the total probability of event B.
If A = Ai in Eq. (1.12), then using (1.16) we have the following important theorem (Bayes theorem):
Example 1.10: In a binary communication system (Figure 1.1), either 0 or 1 is transmitted. Due to channel noise, a 1 can be received as 0 and vice versa. Let m0 and m1 denote the events of transmitting 0 and 1, respectively. Let r0 and r1 denote the events of receiving 0 and 1, respectively. Given P(m0) = 0.5, P(r1/m0) = 0.2 and P(r0/m1) = 0.1, (a) find P(r0) and P(r1). (b) If a 1 was received, what is the probability that a 1 was sent? (c) If a 0 was received, what is the probability that a 0 was sent? (d) Calculate the probability of error, Pe. (e) Calculate the probability (Pc) that the transmitted signal correctly reaches the receiver.
Figure 1.1 A Binary Communication System
Solution:
Using Eq. (1.16), we have



This can also be found as follows:
The outcome of an experiment may either be a real number or be a symbol. An example of the former case is the roll of a die and that of the latter is the toss of a coin. However, for meaningful mathematical analysis, it is always convenient to deal with a real number for each possible outcome. Let S [= {s1, s2, s3, …}] be the sample space of some random experiment. A random or stochastic variable X(si) (or simply X) is a single-valued real function that assigns a real number to each sample point si of S, and the real number is called the value of X(si). Schematic diagram representing a random variable (RV) is shown in Figure 1.2.
Figure 1.2 The Random Variable X(s)
The sample space S is known as the domain of the RV X, and the collection of all values (numbers) assumed by X is called the range or spectrum of the RV X. RV can be either discrete or continuous.
Example 1.11: Two unbiased coins are tossed. Suppose that RV X represents the number of heads that can come up. Find the values taken by X. Also show schematically the mapping of sample points into real numbers on the real axis.
Solution: The sample space S contains four sample points.
Thus,
Table 1.1 illustrates the sample points and the number of heads that can appear (i.e., the values of X). The schematic representation is shown in Figure 1.3.

Figure 1.3 Mapping of Sample Points
Table 1.1 Random variable and its values
Outcome |
Value of X (Number of Heads) |
|---|---|
HH |
2 |
HT |
1 |
TH |
1 |
TT |
0 |
Note that two or more different sample points may assume the same value of X, but two different numbers cannot be assigned to the same sample point.
Consider a discrete RV X that can assume the values x1, x2, x3, …. Suppose that these values are assumed with probabilities given by
This function f(x) is called the probability mass function (PMF), discrete probability distribution or probability function of the discrete RV X. f(x) satisfies the following properties:
Example 1.12: Find the PMF corresponding to the RV X of Example 1.11.
Solution: For the random experiment of tossing two coins, we have
Table 1.2 Tabular Representation of PMF
Thus,
Table 1.2 illustrates the PMF.
The cumulative distribution function (CDF) or briefly the distribution function of a continuous or discrete RV X is given by
The CDF F(x) has the following properties:
The CDF of a discrete RV X is given by
If X assumes only a finite number of values x1, x2, x3, …, then the distribution function can be expressed as follows:

Example 1.13: (a) Find the CDF for the RV X of Example 1.12. (b) Obtain its graph.
Solution:
The distribution function of a continuous RV is represented as follows:
where f(x) satisfies the following conditions:

f(x) is known as the probability density function (PDF) or simply density function.
The following terms are important in the analysis of various probability distributions: Expectation—The expectation or mean of an RV X is defined as follows:
Variance and Standard Deviation—The variance of an RV X is expressed as follows:
Eq. (1.23) is simplified as follows:
The positive square root of the variance (σ) is called the standard deviation of X.
Several probability distributions are frequently used in communication theory. Among them, binomial distribution, Poisson distribution and Gaussian or normal distribution need special mention.
In an experiment of tossing a coin or drawing a card from a pack of 52 cards repeatedly, each toss or selection is called a trial. In some cases, the probability associated with a particular event (such as head on the coin, drawing a heart etc.) is constant in each trial. Such trials are then said to be independent and referred to as Bernoulli trials. Let p be the probability that an event will occur in any single Bernoulli trial (probability of success). Therefore, q = 1 − p is the probability that the event will fail to occur (probability of failure). The probability that the event will occur exactly k times in n trials (i.e. k successes and (n − k) failures) is given by the probability function
where
is known as the binomial coefficient. The corresponding RV X is called a binomial RV and the CDF of X is
The mean and variance of the binomial RV X are, respectively,
The binomial distribution finds application in digital transmission, where X stands for the number of errors in a message of n digits.
Example 1.14: A binary source generates digits 1 and 0 randomly with probabilities 0.7 and 0.3, respectively. What is the probability that three digits of a five-digit sequence will be 1?
Solution:

Let X be a discrete RV. It is called a Poisson RV if its PMF is expressed as
where α is a positive constant. The above distribution is called the Poisson distribution. The corresponding CDF is given by
The mean and variance in this case are, respectively,
Poisson distribution finds application in monitoring the number of telephone calls coming at a switching centre during different intervals of time. Binomial distribution fails to solve the problem of the transmission of many data bits, where the error rate is low. In such a situation Poisson distribution proves to be effective.
The PDF of a Gaussian or normal RV X is given by
where μ and σ are the mean and standard deviation, respectively. The corresponding CDF of X is as follows:
Eq. (1.30) cannot be evaluated analytically and is solved numerically. We define a function Q(z) such that
Thus, Eq. (1.30) is rewritten as follows:
Q(z) is known as the complementary error function. Figure 1.5 illustrates a Gaussian distribution.
Figure 1.5 Gaussian Distribution
The normal distribution is useful in describing the random phenomena in nature. The sum of a large number of independent random variables, under certain conditions, can also be approximated by a normal distribution (central-limit theorem).
In the context of communication theory, an RV that is a function of time is called a random process or a stochastic process. Thus, a random process is collection of an infinite number of RVs. Any random signal such as noise is characterized by random process. In order to determine the statistical behaviour of a random process, we might proceed in either of the two ways: We might perform repeated measurement of the same random process or we might take simultaneous measurements of a very large number of identical random processes.
Let us consider a random process such as a noise waveform. To obtain the statistical properties of the noise, we might repeatedly measure the output noise voltage of a single noise source or we might make simultaneous measurements of the output of a very large collection of statistically identical noise sources. The collection of all possible waveforms is known as the ensemble (X(t, s) or simply X(t)), and a waveform in this collection is a sample function or ensemble member or realization of the process.
Stationary Random Process—A random process is called stationary random process if the statistical properties of the process remain unaltered with time.
Strict-Sense Stationary—A random process is said to be strict-sense stationary if its probability density does not change with shift of origin, i.e.,
Wide-Sense Stationary—A random process is called wide-sense stationary if its mean is a constant and autocorrelation depends only on time difference.
Thus,
and
Ergodic Process—A stationary process is known as ergodic process if the ensemble average is equal to the time average. The time-averaged mean of a sample function x(t) of a random process X(t) is given by
Thus, in this case,

The process which is not stationary is termed as non-stationary.
The term noise encompasses all forms of unwanted electrical signal that corrupt the message signal during its transmission or processing in a communication system. Along with other basic factors, it sets limits on the rate of communication. However, it is often possible to suppress or even eliminate the effects of noise by intelligent circuits design. Thus, the study of fundamental sources of noise and their characteristics are very much essential in the context of communication.
Noise arises in a variety of ways. Potential sources of random noise can be broadly classified as external noise and internal noise.
Noise arising from these sources is difficult to suppress and often the only solution is to reposition the receiving antenna to minimize the received noise, while ensuring that reception of the desired signal is not seriously impaired.
The internal noise is produced by the active and passive electronic components within the communication circuit. It arises due to
Some of the main constituents of internal noise are thermal noise, shot noise, partition noise and flicker noise.
In a conductor, the free electrons are thermally agitated because heat exchange takes place between the conductor and its surroundings. These free electrons exhibit random motion as a result of their collisions with lattice structure. Consequently, there exists a random variation in the electron density throughout the conductor, resulting in a randomly fluctuating voltage across the ends of the conductor. This is the most important class of noise in electronic circuits and known as thermal noise or Johnson noise or resistor noise.
Thermal noise can be best be described by a zero mean Gaussian process and has a flat type of power spectral density over a wide range of frequencies.
Shot noise appears in all amplifying and active devices as a random fluctuation superimposed on the steady current crossing a potential barrier. The effect occurs because electrical current is not a continuous but discrete flow of charge carriers. In vacuum tubes, shot noise is caused by random emission of electrons from cathode. In semiconductor devices, it is generated due to the random variation in the diffusion of minority carriers (holes and electrons) and random generation and recombination of electron–hole pairs. As lead shots from a gun strike a target, in a similar way, electrons from the cathode strike the anode plate in a vacuum tube; hence, the name shot noise. Although shot noise is always present, its effect is significant in the amplifying devices.
The fluctuating component of current (in(t)) which wiggles around the mean value Idc is the shot noise. Thus, the total current i(t) is expressed as follows:

Like thermal noise, shot noise also has flat type of power spectrum (power spectral density) except in the high microwave frequency range. For diodes, the rms shot noise current is given by

where Idc is the direct diode current, q is the electronic charge (= 1.6 × 1019 C) and B is the bandwidth of the system.
Partition noise is generated in multi-grid tubes (tetrode, pentode, etc.) where current divides between two or more electrodes due to random fluctuations of electrons emitted from cathode among various electrodes. In this respect, a diode would be less noisy than a transistor. Hence, the input stage of microwave receivers is often a diode circuit. Very recently, GaAs field effect transistors have been developed for low-noise microwave amplification. The spectrum for partition noise is also flat.
At low audio frequencies (less than few kilohertz), another type of noise arises in vacuum tubes and semiconductor devices, the power spectral density of which is inversely proportional to frequency. This noise is known as flicker noise or 1/f noise. In vacuum tubes, it stems due to gradual changes in the oxide structure of oxide-coated cathodes and due to the migration of impurity ions through the oxide. In case of semiconductor devices, it is produced due to contaminants and defects in the crystal structure.
Problem 1.1: What is the probability that either a 1 or 6 will appear in the experiment involving the rolling of a die?
Solution: Since each of the six outcomes is equally likely in a large number of trials, each outcome has a probability of 1/6.
Thus, the probability of occurrence of either 1 or 6 is given by
Problem 1.2: Consider an experiment of drawing two cards from a deck of cards without replacing the first card drawn. Determine the probability of obtaining two red aces in two draws.
Solution: Let A and B be two events ‘red ace in the first draw’ and ‘red ace in the second draw’, respectively.
We know that P(A ∩ B) = P(A)P(B/A)
The relative frequency of 
P(B/A) is the probability of drawing a red ace in the second draw provided that the outcome of the first draw was a red ace.
Problem 1.3: The random variable X assumes the values 0 and 1 with probabilities p and q = 1 − p, respectively. Find the mean and the variance of X.
Solution:

Problem 1.4: A fair die is rolled 5 times. If 1 or 6 appears, it is called success. Determine (a) the probability of two successes and (b) the mean and standard deviation for the number of successes.
Solution: We apply binomial distribution.


Problem 1.5: When n is very large (n >> k) and p very small (p << 1), prove that the binomial distribution can be approximated by the following Poisson distribution:
Solution: From binomial distribution, we write
Since n >> k and p << 1, then
and
Thus,
Substituting these values into Eq. (1.33) we get
Ans. (c)
Ans. (a)
Ans. (b)
Ans. (b)
Ans. (d)
The value of P(A ∪ B) is
Ans. (a)
Ans. (a)
Ans. (d)
Ans. (c)
Ans. (d)
Ans. (c)
Var (X) is
Ans. (a)
Ans. (b)
Ans. (a)
Ans. (b)



are mutually independent.
Find the mean and variance of X.
