Distribution density of the sum of two uniformly distributed quantities. The law of distribution of the sum of two random variables

THEME 3

concept of distribution function

mathematical expectation and variance

uniform (rectangular) distribution

normal (Gaussian) distribution

Distribution

t- Student's distribution

F- distribution

distribution of the sum of two random independent variables

example: distribution of the sum of two independent

uniformly distributed quantities

random variable transformation

example: distribution of a harmonic wave

with random phase

central limit theorem

moments of a random variable and their properties

PURPOSE OF THE CYCLE

LECTURES:

REPORT INITIAL INFORMATION ABOUT THE MOST IMPORTANT DISTRIBUTION FUNCTIONS AND THEIR PROPERTIES

DISTRIBUTION FUNCTIONS

Let be x(k) is some random variable. Then for any fixed value x a random event x(k) x defined as the set of all possible outcomes k such that x(k) x. In terms of the original probability measure given on the sample space, distribution functionP(x) defined as the probability assigned to a set of points k x(k) x. Note that the set of points k satisfying the inequality x(k) x, is a subset of the set of points that satisfy the inequality x(k). Formally

It's obvious that

If the range of values ​​of the random variable is continuous, which is assumed below, then probability density(one-dimensional) p(x) is determined by the differential relation

(4)

Hence,

(6)

In order to be able to consider discrete cases, it is necessary to admit the presence of delta functions in the composition of the probability density.

EXPECTED VALUE

Let the random variable x(k) takes values ​​from the range from -  to + . Mean(otherwise, expected value or expected value) x(k) is calculated using the corresponding passage to the limit in the sum of products of values x(k) on the probability of these events occurring:

(8)

where E- mathematical expectation of the expression in square brackets by index k. The mathematical expectation of a real single-valued continuous function is defined similarly g(x) from a random variable x(k)

(9)

where p(x)- probability density of a random variable x(k). In particular, taking g(x)=x, we get mean square x(k) :

(10)

Dispersionx(k) defined as the mean square of the difference x(k) and its average value,

i.e. in this case g(x)= and

A-priory, standard deviation random variable x(k), denoted , is the positive value of the square root of the variance. The standard deviation is measured in the same units as the mean.

MOST IMPORTANT DISTRIBUTION FUNCTIONS

UNIFORM (RECTANGULAR) DISTRIBUTION.

Let us assume that the experiment consists in a random selection of a point from the interval [ a,b] , including its endpoints. In this example, as the value of a random variable x(k) you can take the numeric value of the selected point. The corresponding distribution function has the form

Therefore, the probability density is given by the formula

In this example, the calculation of the mean and variance using formulas (9) and (11) gives

NORMAL (GAUSSIAN) DISTRIBUTION

, - arithmetic mean, - RMS.

The value of z corresponding to the probability P(z)=1-, i.e.

CHI - SQUARE DISTRIBUTION

Let be - n independent random variables, each of which has a normal distribution with zero mean and unit variance.

Chi-squared random variable with n degrees of freedom.

probability density .

DF: 100 - percentage points - distributions are denoted by , i.e.

mean and variance are equal

t - STUDENT DISTRIBUTIONS

y, z are independent random variables; y - has - distribution, z - normally distributed with zero mean and unit variance.

value - has t- Student's distribution with n degrees of freedom

DF: 100 - percentage point t - distribution is indicated

Mean and variance are equal

F - DISTRIBUTION

Independent random variables; has - distribution with degrees of freedom; distribution with degrees of freedom. Random value:

,

F is a distributed random variable with and degrees of freedom.

,

DF: 100 - percentage point:

The mean and variance are equal:

DISTRIBUTION OF THE AMOUNT

TWO RANDOM VARIABLES

Let be x(k) and y(k) are random variables having a joint probability density p(x,y). Find the probability density of the sum of random variables

At a fixed x we have y=z–x. So

At a fixed z values x run the interval from – to +. So

(37)

whence it can be seen that in order to calculate the desired density of the sum, one must know the original joint probability density. If a x(k) and y(k) are independent random variables having densities and, respectively, then and

(38)

EXAMPLE: THE SUM OF TWO INDEPENDENT, UNIFORMLY DISTRIBUTED RANDOM VARIABLES.

Let two random independent variables have densities of the form

In other cases Let us find the probability density p(z) of their sum z= x+ y.

Probability Density for i.e. for Hence, x less than z. In addition, is not equal to zero for By formula (38), we find that

Illustration:

The probability density of the sum of two independent, uniformly distributed random variables.

RANDOM CONVERSION

VALUES

Let be x(t)- random variable with probability density p(x), let it go g(x) is a single-valued real continuous function of x. Consider first the case when the inverse function x(g) is also a single-valued continuous function of g. Probability Density p(g), corresponding to a random variable g(x(k)) = g(k), can be determined from the probability density p(x) random variable x(k) and derivative dg/dx under the assumption that the derivative exists and is different from zero, namely:

(12)

Therefore, in the limit dg/dx#0

(13)

Using this formula, follows on its right side instead of a variable x substitute the appropriate value g.

Consider now the case when the inverse function x(g) is valid n-valued function of g, where n is an integer and all n values ​​are equally probable. Then

(14)

EXAMPLE:

DISTRIBUTION OF THE HARMONIC FUNCTION.

Harmonic function with fixed amplitude X and frequency f will be a random variable if its initial phase angle = (k)- random value. In particular, let t fixed and equal t o, and let the harmonic random variable have the form

Let's pretend that (k) has a uniform probability density p() kind

Find the probability density p(x) random variable x(k).

In this example, the direct function x() unambiguously, and the inverse function (x) ambiguous.

In practice, it often becomes necessary to find the distribution law for the sum of random variables.

Let there be a system (X b X 2) two continuous s. in. and their sum

Let us find the distribution density c. in. U. In accordance with the general solution of the previous paragraph, we find the region of the plane where x + x 2 (Fig. 9.4.1):

Differentiating this expression with respect to y, we obtain an ap. random variable Y \u003d X + X 2:

Since the function φ (x b x 2) = Xj + x 2 is symmetric with respect to its arguments, then

If with. in. X and X 2 are independent, then formulas (9.4.2) and (9.4.3) take the form:


In the case when independent c. in. x x and X 2, talk about the composition of distribution laws. Produce composition two distribution laws - this means finding the distribution law for the sum of two independent c. c., distributed according to these laws. The symbolic notation is used to designate the composition of distribution laws

which is essentially denoted by formulas (9.4.4) or (9.4.5).

Example 1. The work of two technical devices (TD) is considered. First, TU works after its failure (failure) is included in the operation of TU 2. Uptime TU TU TU 2 - x x and X 2 - are independent and distributed according to exponential laws with parameters A,1 and X 2 . Therefore, the time Y trouble-free operation of the TU, consisting of TU! and TU 2 will be determined by the formula

It is required to find a p.r. random variable Y, i.e., the composition of two exponential laws with parameters and X 2 .

Decision. By formula (9.4.4) we get (y > 0)


If there is a composition of two exponential laws with the same parameters (?c = X 2 = Y), then in the expression (9.4.8) an uncertainty of type 0/0 is obtained, expanding which, we get:

Comparing this expression with expression (6.4.8), we are convinced that the composition of two identical exponential laws (?c = X 2 = x) is the second-order Erlang law (9.4.9). When composing two exponential laws with different parameters x x and A-2 get second-order generalized Erlang law (9.4.8). ?

Problem 1. The law of distribution of the difference of two s. in. System with. in. (X and X 2) has a joint r.p./(x x x 2). Find a p.r. their differences Y=X - X 2 .

Decision. For the system with in. (X b - X 2) etc. will be / (x b - x 2), i.e. we replaced the difference with the sum. Therefore, a.r. random variable U will have the form (see (9.4.2), (9.4.3)):

If a with. in. X x iX 2 independent, then

Example 2. Find a f.r. the difference of two independent exponentially distributed s. in. with parameters x x and X 2 .

Decision. According to the formula (9.4.11) we get

Rice. 9.4.2 Rice. 9.4.3

Figure 9.4.2 shows a p. g(y). If we consider the difference of two independent exponentially distributed s. in. with the same settings (A-i= X 2 = BUT,), then g(y) \u003d / 2 - already familiar

Laplace's law (Fig. 9.4.3). ?

Example 3. Find the distribution law for the sum of two independent c. in. X and X 2, distributed according to the Poisson law with parameters a x and a 2 .

Decision. Find the probability of an event (X x + X 2 = t) (t = 0, 1,



Therefore, s. in. Y= X x + X 2 distributed according to the Poisson law with the parameter a x2) - a x + a 2. ?

Example 4. Find the distribution law for the sum of two independent c. in. x x and X 2, distributed according to binomial laws with parameters p x ri p 2 , p respectively.

Decision. Imagine with. in. x x as:

where X 1) - event indicator BUT wu "th experience:

Distribution range with. in. X,- has the form


We will make a similar representation for s. in. X 2: where X] 2) - event indicator BUT in y"-th experience:


Hence,

where is X? 1)+(2) if the event indicator BUT:

Thus, we have shown that in. Father-in-law amount (u + n 2) event indicators BUT, whence it follows that s. in. ^distributed according to the binomial law with parameters ( n x + n 2), p.

Note that if the probabilities R in different series of experiments are different, then as a result of adding two independent s. c., distributed according to binomial laws, it turns out c. c., distributed not according to the binomial law. ?

Examples 3 and 4 are easily generalized to an arbitrary number of terms. When composing Poisson's laws with parameters a b a 2 , ..., a t Poisson's law is again obtained with the parameter a (t) \u003d a x + a 2 + ... + and t.

When composing binomial laws with parameters (n r); (i 2 , R) , (n t, p) again we get the binomial law with parameters (“(“), R), where n (t) \u003d u + n 2 + ... + etc.

We have proved important properties of Poisson's law and the binomial law: the "stability property". The distribution law is called sustainable, if the composition of two laws of the same type results in a law of the same type (only the parameters of this law differ). In Subsection 9.7 we will show that the normal law has the same stability property.

An extremely important object of probability theory is the sum of independent random variables. It is the study of the distribution of sums of independent random variables that laid the foundation for the development of analytical methods of probability theory.

Distribution of the sum of independent random variables

In this section, we will obtain a general formula that allows us to calculate the distribution function of the sum of independent random variables, and consider several examples.

Distribution of the sum of two independent random variables. Convolution Formula

independent random variables with distribution functions

respectively

Then the distribution function F sums of random variables

can be calculated using the following formula ( convolution formula)

To prove this, we use Fubini's theorem.

The second part of the formula is proved similarly.

Distribution density of the sum of two independent random variables

If the distributions of both random variables have densities, then the density of the sum of these random variables can be calculated by the formula

If the distribution of a random variable (or ) has a density, then the density of the sum of these random variables can be calculated by the formula

To prove these assertions, it suffices to use the definition of density.

Multiple convolutions

The calculation of the sum of a finite number of independent random variables is performed using the sequential application of the convolution formula. Sum distribution function k independent identically distributed random variables with a distribution function F

called k–fold convolution of the distribution function F and denoted

Examples of calculating the distribution of sums of independent random variables

In this paragraph, examples of situations are given, when summing random variables, the form of the distribution is preserved. The proofs are exercises on summation and calculation of integrals.

Sums of independent random variables. Normal distribution

Sums of independent random variables. Binomial distribution

Sums of independent random variables. Poisson distribution

Sums of independent random variables. Gamma distribution

Poisson process

a sequence of independent identically distributed random variables having an exponential distribution with parameter



Random sequence of points

on the non-negative semi-axis is called Poisson (point) process.

Let us calculate the distribution of the number of points

Poisson process in the interval (0,t)

equivalents, so

But the distribution of the random variable

is an Erlang distribution of order k, so

Thus, the distribution of the number of points of the Poisson process in the interval (o,t) is a Poisson distribution with the parameter

The Poisson process is used to simulate the moments of occurrence of random events - the process of radioactive decay, the moments of receipt of calls to the telephone exchange, the moments of the appearance of customers in the service system, the moments of equipment failure.

Let us use the above general method to solve one problem, namely, to find the distribution law for the sum of two random variables. There is a system of two random variables (X,Y) with distribution density f(x,y).

Consider the sum of random variables X and Y: and find the law of distribution of the value Z. To do this, we construct a line on the xOy plane, the equation of which (Fig. 6.3.1). This is a straight line cutting off segments equal to z on the axes. Straight divides the xy plane into two parts; to the right and above ; left and below

Region D in this case is the lower left part of the xOy plane, shaded in Fig. 6.3.1. According to formula (6.3.2) we have:

This is the general formula for the distribution density of the sum of two random variables.

For reasons of symmetry of the problem with respect to X and Y, we can write another version of the same formula:

It is required to produce a composition of these laws, i.e., to find the distribution law of the quantity: .

We apply the general formula for the composition of distribution laws:

Substituting these expressions into the formula we have already encountered

and this is nothing but a normal law with a dispersion center

The same conclusion can be reached much more easily with the help of the following qualitative reasoning.

Without opening the brackets and without performing transformations in the integrand (6.3.3), we immediately come to the conclusion that the exponent is a square trinomial with respect to x of the form

where the value of z is not included in the coefficient A at all, it is included in the coefficient B in the first degree, and the coefficient C is included in the square. With this in mind and applying formula (6.3.4), we conclude that g(z) is an exponential function, the exponent of which is a square trinomial with respect to z, and the distribution density; of this kind corresponds to the normal law. Thus, we; we come to a purely qualitative conclusion: the law of distribution of z must be normal. To find the parameters of this law - and - use the theorem of addition of mathematical expectations and the theorem of addition of variances. According to the addition theorem of mathematical expectations . According to the variance addition theorem or whence formula (6.3.7) follows.

Passing from root-mean-square deviations to probable deviations proportional to them, we get:
.

Thus, we have come to the following rule: when normal laws are composed, a normal law is again obtained, and the mathematical expectations and variances (or squared probable deviations) are summed up.

The composition rule for normal laws can be generalized to the case of an arbitrary number of independent random variables.

If there are n independent random variables: subject to normal laws with dispersion centers and standard deviations , then the value is also subject to the normal law with parameters

If the system of random variables (X, Y) is distributed according to the normal law, but the quantities X, Y are dependent, then it is easy to prove, just as before, based on the general formula (6.3.1), that the distribution law of the quantity is also a normal law. Scattering centers still add algebraically, but for standard deviations the rule becomes more complicated: , where, r is the correlation coefficient of X and Y values.

When adding several dependent random variables that in their totality obey the normal law, the distribution law of the sum also turns out to be normal with parameters

where is the correlation coefficient of the quantities X i , X j , and the summation extends to all different pairwise combinations of the quantities .

We have seen a very important property of the normal law: when normal laws are combined, one again obtains a normal law. This is the so-called "stability property". A distribution law is said to be stable if, by composing two laws of this type, a law of the same type is again obtained. We have shown above that the normal law is stable. Very few distribution laws have the property of stability. The law of uniform density is unstable: when composing two laws of uniform density in sections from 0 to 1, we obtained Simpson's law.

The stability of a normal law is one of the essential conditions for its wide application in practice. However, the property of stability, in addition to the normal one, is also possessed by some other distribution laws. A feature of the normal law is that when a sufficiently large number of practically arbitrary distribution laws are composed, the total law turns out to be arbitrarily close to the normal one, regardless of what the distribution laws of the terms were. This can be illustrated, for example, by composing the composition of three laws of uniform density in sections from 0 to 1. The resulting distribution law g(z) is shown in fig. 6.3.1. As can be seen from the drawing, the graph of the function g(z) is very similar to the graph of the normal law.

Let there be a system of two random variables X and Y, whose joint distribution is known. The task is to find the distribution of a random variable . As examples of SV Z you can bring profit from two enterprises; the number of voters who voted in a certain way from two different precincts; the sum of the points on the two dice.

1. The case of two DSVs. Whatever values ​​the discrete CVs take (in the form of a finite decimal fraction, with different steps), the situation can almost always be reduced to the following special case. Quantities X and Y can only take integer values, i.e. where . If initially they were decimal fractions, then they can be made integers by multiplying by 10 k. And the missing values ​​between the highs and lows can be assigned zero probabilities. Let the joint probability distribution be known. Then, if we number the rows and columns of the matrix according to the rules: , then the probability of the sum is:

The elements of the matrix are added along one of the diagonals.

2. The case of two NSWs. Let the joint distribution density be known. Then the distribution density of the sum:

If a X and Y independent, i.e. , then

Example 1 X , Y– independent, uniformly distributed SW:

Let's find the distribution density of the random variable .

It's obvious that ,

SW Z can take values ​​in the interval ( c+d; a+b), but not for all x. outside of this interval. On the coordinate plane ( x, z) the range of possible values ​​of the quantity z is a parallelogram with sides x=with; x=a; z=x+d; z=x+b. In the formula for the limits of integration will be c and a. However, due to the fact that in the replacement y=z-x, for some values z function . For example, if c , then at z=x+c and any x will have: . Therefore, the calculation of the integral should be carried out separately for different areas of change in the value z, in each of which the limits of integration will be different, but for all x and z. We will do this for the special case when a+d< b+c . Let us consider three different regions of change in the quantity z and for each of them we find .

1) c+d ≤ z ≤ a+d. Then

2) a+d ≤ z ≤ b+c. Then

3) b+c ≤ z ≤ a+b. Then

This distribution is called Simpson's law. Figures 8, 9 show graphs of SW distribution density at with=0, d=0.