120 views

Uploaded by Abd El-Hai Mohammed

- Normal Distribution - Wikipedia, The Free Encyclopedia
- 320
- 2-Way-TOS-Reading and Writing -.xlsx
- Statistics Problems 2 Chapter 9
- Time History Ground Motion Scaling in Etabs
- x Ray Diffraction and Nmr Data for the Study of the Location of Idebenone and Idebenol in Model Membranes
- An Analysis of Pressure Fluctuations in a CFB of Heavy Minerals
- Normal Distribution
- hw4
- 100524
- Shecdiling.pdf
- 2.2
- The Measurement of Cholinesterase Activities as a Biomarker
- Project 2010
- rr311801-probability-and-statistics
- qch11-12-13
- 25101-720_UE Radio Transmission and Reception (FDD)
- Biological_Variation_update_ed.ppt
- 2013_Ch.13_Notes-2
- Chapter 6

You are on page 1of 65

This packet is transmitted over a communications channel where the probability of error for each bit is 10-3. What is the probability that the packet is received error-free?

Problem 8.1

Solution Recognizing that the number of errors has a binomial distribution over the sequence of 200 bits, let x represent the number of errors with p = 0.001 and n = 200. Then the probability of no errors is

P x = 0 = (1 p )

n 200

Excerpts from this work may be reproduced by instructors for distribution on a not-for-profit basis for testing or instructional purposes only to students enrolled in courses for which the textbook has been adopted. Any other reproduction or translation of this work beyond that permitted by Sections 107 or 108 of the 1976 United States Copyright Act without the permission of the copyright owner is unlawful.

page8-1

Problem 8.2 Suppose the packet of the Problem 8.1 includes an error-correcting code that can correct up to three errors located anywhere in the packet. What is the probability that a particular packet is received in error in this case? Solution The probability of a packet error is equal to the probability of more than three bit errors. This is equivalent to 1 minus the probability of 0, 1, 2, or 3 errors:

1 P[x 3] = 1 (P[x = 0] + P[x = 1] + P[x = 2] + P[x = 3]) n n n = 1 (1 p) n p(1 p )n1 p 2 (1 p )n 2 p 3 (1 p )n 3 3 2 1 n(n 1) 2 n(n 1)(n 2) 3 p (1 p ) + p = 1 (1 p )n3 (1 p )3 + np(1 p )2 + 2 6 = 5.5 10 5

Excerpts from this work may be reproduced by instructors for distribution on a not-for-profit basis for testing or instructional purposes only to students enrolled in courses for which the textbook has been adopted. Any other reproduction or translation of this work beyond that permitted by Sections 107 or 108 of the 1976 United States Copyright Act without the permission of the copyright owner is unlawful.

page8-2

Problem 8.3 Continuing with Example 8.6, find the following conditional probabilities: P[X=0|Y=1]

and P[X =1|Y=0].

Solution

From Bayes Rule

P X = 0Y = 1 = P Y = 1 X = 0 P[ X = 0 ]

P X = 1Y = 0 = =

Excerpts from this work may be reproduced by instructors for distribution on a not-for-profit basis for testing or instructional purposes only to students enrolled in courses for which the textbook has been adopted. Any other reproduction or translation of this work beyond that permitted by Sections 107 or 108 of the 1976 United States Copyright Act without the permission of the copyright owner is unlawful.

page8-3

Problem 8.4 Consider a binary symmetric channel for which the conditional probability of error p = 10-4, and symbols 0 and 1 occur with equal probability. Calculate the following probabilities:

a) The probability of receiving symbol 0. b) The probability of receiving symbol 1. c) The probability that symbol 0 was sent, given that symbol 0 is received d) The probability that symbol 1 was sent, given that symbol 0 is received.

= .9999 1 + .0001 1 2 2 =1 2

(b)

P[Y = 1] = 1 P[Y = 0] =1 2

(c)

From Eq.(8.30)

P [X = 0 Y = 0 ] = =

(1 p ) p0 (1 p ) p0 + pp1

4 1 4 1 2 4

(1 10 ) (1 10 ) + 10

2

4 1

= 1 10

(d)

P [X = 1Y = 0 ] = = pp1 pp1 + (1 p ) p0 10 4 10 4 12 1 + (1 10 4 ) 1 2 2

= 10 4

page8-4

Problem 8.5 Determine the mean and variance of a random variable that is uniformly distributed

between a and b.

= E[X ] =

= x

a b

xf

b

( x)dx

1 dx ba

The variance is given by

E (X ) =

2

] (x )

b

f X ( x)dx

ba 3 3 1 (b ) (a ) = ba 3 3

a

(x )2 dx

If we substitute =

E (X ) =

2

b+a then 2

(b a )2 =

12

3 3 1 (b a ) (a b ) b a 24 24

page8-5

Problem 8.6 Let X be a random variable and let Y = (X-X)/X. What is the mean and variance of the random variable Y? Solution

X X E[X ] X 0 E[Y ] = E = =0 = X X X

E(Y Y ) = E Y

2

[ ]

2

E( X X )

X X = E X

2

X2

X2 = =1 X2

page8-6

Problem 8.7 What is the probability density function of the random variable Y of Example 8.8? Sketch

this density function.

Solution From Example 8.8, the distribution of Y is 0 1 2 2 cos ( y ) FY ( y ) = 2 1 Thus, the density of Y is given by

0 dFY ( y ) 1 = 2 dy 1 y 0

fY(y)

y

-1 1

page8-7

Problem 8.8 Show that the mean and variance of a Gaussian random variable X with the density 2 function given by Eq. (8.48) are and X .

X

E[X ] X =

(x X ) exp (x X )2 dx

2 X 2 X

2

E[ X ] X =

y 2 X

2 exp y 2 dy 2 X

=0

since integrand has odd symmetry. This implies E[X ] = X . With this result

Var( X ) = E( x X ) = In this case let

y= x X

2

(x X )2 exp (x X )2 dx

2 X 2 X

2

2

y2 y2 exp dy 2 2

y2 dv = y exp 2 dy . Then

Continued on next slide

page8-8

Var ( X ) = X

2

( ) y exp y 2

2

2

y2 1 2 + X exp dy 2 2 2

= 0 + X 1 =X

2

where the second integral is one since it is integral of the normalized Gaussian probability density.

page8-9

2 Problem 8.9 Show that for a Gaussian random variable X with mean X and variance X the

E[Y ] = =0

1 2

2 y exp y dy 2

by the odd symmetry of the integrand. If E[Y] = 0, then from the definition of Y, E[X] = X. In a similar fashion

EY2 =

[ ]

1 2

y2 y 2 exp 2 dy

y2 () y 1 = exp + 2 2 2 =1

2

2 exp y dy 2

x X E X

2 and hence E( x X ) = X 2

=1

page8-10

Problem 8.10 Determine the mean and variance of the sum of five independent uniformly-distributed random variables on the interval from -1 to +1. Solution

Let Xi be the individual uniformly distributed random variables for i = 1,..,5, and let Y be the random variable representing the sum:

Y = Xi

i =1 5

Since Xi has zero mean and Var(Xi) = 1/3 (see Problem 8.5), we have

E[Y ] = E[X i ] = 0

i =1 5

and

Var (Y ) = E (Y Y ) = E Y 2

2

= E ( X i )

5

] [ ]

[ ]

i j

= E X i2 + E X i X j

i =1

[ ]

Var(Y ) = 5 1 + E[ X i ]E X j 3 = 5 +0 3 5 = 3

( )

[ ]

page8-11

X (t , ) = A cos(2ft + )

where A and f are constants, and is uniformly distributed over the interval 0 to 2. Is X stationary to the first order?

Solution Denote

Y = X (t1 , ) = A cos(2ft1 + )

for any t1. From Problem 8.7, the distribution of Y and therefore of X for any t1 is

0 2 2 cos 1 ( y / A) FX (t1 ) ( y ) = 2 1

page8-12

Problem 8.12 Show that a random process that is stationary to the second order is also stationary to the

first order.

Solution

Then,

FX ( t1 ) X (t2 ) ( x1 , ) = FX (t1 ) (x1 ) = FX ( t1 + ) X ( t2 + ) ( x1 , ) = FX ( t1 + ) ( x1 )

page8-13

X (t ) = A cos(2ft )

where A is uniformly distributed between 0 and 1, and f is constant. Determine the autocorrelation function of X. Is X wide-sense stationary?

Solution

E[X (t1 )X (t 2 )] = E A 2 cos(2ft1 ) cos(2ft 2 )

2 1

[ ] = E[A ][cos(2f (t

1

t 2 )) + cos 2f (t1 + t 2 )]

x3 1 E A = x dx = = 0 3 0 3

2 1 2

[ ]

Since the autocorrelation function depends on t1 + t 2 as well as t1 t 2 , the process is not wide-sense stationary.

page8-14

Yn = 0 Z n + 1 Z n 1

where {Zn} is a random process with autocorrelation function RZ (n) =

2

Solution

We implicitly assume that Zn is stationary and has a constant mean Z. Then the mean of Yn is given by

The autocorrelation of Y is given by

E[YnYm ] = E[( 0 Z n + 0 Z n1 )( 0 Z m + 1Z m1 )]

2 = 0 E[Z n Z m ] + 1 0 E[Z n Z m1 ] + 01E[Z n1 Z m ] + 12 E[Z m1 Z n1 ]

2 = 0 2 (n m ) + 1 0 2 (m 1 n ) + 01 2 (n 1 m ) + 12 (m 1 (n 1)) 2 = 0 + 12 2 (n m ) + 01 2 [ (n m 1) + (m n 1)]

Since the autocorrelation only depends on the time difference n-m, the process is widesense stationary with

2 RY (n) = 0 + 12 2 (n) + 01 2 ( (n 1) + (n + 1) )

page8-15

Problem 8.15 For the discrete-time process of Problem 8.14, use the discrete Fourier transform to

approximate the corresponding spectrum. That is,

S Y (k ) = RY (n)W kn

n=0

N 1

If the sampling in the time domain is at n/Ts where n = 0, 1, 2, , N-1. What frequency does k correspond to?

Solution Let

0 = ( 02 + 12 ) 2 and 1 = 01 2 . Then

SY (k ) = [ 0 (n ) + 1 ( (n 1) + (n + 1))] W kn

n =0 N 1

= 0W 0 + 1 W k + W + k

+ j 2k 2 jNk e = 0 + 1 +e N 2k = 0 + 2 1 cos N

kf s 1 where f S = . TS N

page8-16

Problem 8.16 Is the discrete-time process {Yn: n = 1,2,} defined by: Y0 = 0 and

Yn +1 = Yn + Wn ,

a Gaussian process, if Wn is Gaussian?

Solution (Proof by mathematical induction.) The first term Y1 = Y0 + W0 is Gaussian since Y0 = 0 and W0 are Gaussian. The second term Y2 = Y1 + W1 is Gaussian since Y1 and W1 are Gaussian. Assume Yn is Gaussian. Then Yn +1 = Yn + Wn is Gaussian since Yn and Wn are both Gaussian.

page8-18

Problem 8.17 A discrete-time white noise process {Wn} has an autocorrelation function given by RW(n) = N0(n). (a) Using the discrete Fourier transform, determine the power spectral density of {Wn}. (b) The white noise process is passed through a discrete-time filter having a discretefrequency response

H (k ) = 1 (W k ) N 1 W k

where, for a N-point discrete Fourier transform, W = exp{j2/N}. What is the spectrum of the filter output?

S (k ) = R(n ) W nk

n =0 N 1

= N 0 (n )W nk

n =0

N 1

= N0

SY (k ) = H (k ) S (k )

2

1 (W k ) N = N0 1 W k

page8-19

Problem 8.18 Consider a deck of 52 cards, divided into four different suits, with 13 cards in each suit ranging from the two up through the ace. Assume that all the cards are equally likely to be drawn.

(a) Suppose that a single card is drawn from a full deck. What is the probability that this card is the ace of diamonds? What is the probability that the single card drawn is an ace of any one of the four suits? (b) Suppose that two cards are drawn from the full deck. What is the probability that the cards drawn are an ace and a king, not necessarily the same suit? What if they are of the same suit?

Solution (a)

1 52

P[Ace and king ] = P[Ace on first draw ]P[King on second] + P[King on first draw ]P[Ace on seco 1 4 1 4 = + 13 51 13 51 8 = 663 1 1 1 1 P[Ace and king of same suit ] = + 13 51 13 51 1 = 663

page8-20

Problem 8.19

Suppose a player has one red die and one white die. How many outcomes are possible in the random experiment of tossing the two dice? Suppose the dice are indistinguishable, how many outcomes are possible?

Solution The number of possible outcomes is 6 6 = 36 , if distinguishable. If the die are indistinguishable then the outcomes are (11) (12)(16) (22)(23)(26) (33)(34)(36) (44)(45)(46) (55)(56) (66) And the number of possible outcomes are 21.

page8-21

(a) What is the probability of throwing a red 5 and a white 2? (b) If the dice are indistinguishable, what is the probability of throwing a sum of 7? If they are distinguishable, what is this probability?

1 1 1 = 6 6 36

(b) The probability of the sum does not depend upon whether the die are distinguishable or not. If we consider the distinguishable case the possible outcomes are (1,6), (2,5), (3,4), (4,3), (5,2), and (6,1) so 6 1 P[sum of 7] = = 36 6

page8-22

Problem 8.21 Consider a random variable X that is uniformly distributed between the values of 0 and 1 with probability takes on the value 1 with probability and is uniformly distributed between values 1 and 2 with probability . Determine the distribution function of the random variable X. Solution

0 x 4 FX ( x) = 1 2 1 + 1 ( x 1) 2 2 1 x0 0 < x <1 x =1 1< x 2 x>2

page8-24

Problem 8.22 Consider a random variable X defined by the double-exponential density where a and b are constants.

f X ( x) = a exp( b x )

< x<

(a) Determine the relationship between a and b so that fX(x) is a probability density function. (b) Determine the corresponding distribution function FX(x). (c) Find the probability that the random variable X lies between 1 and 2.

Solution (a)

0

2a exp( bx ) = 1 0 b b = 2a

2a = 1 or b

(b)

FX ( x ) = a exp( b s )ds

x

a x < x <0 exp( b( s )) b = 1 + a exp( bs ) x 0 < x < 2 0 b a < x <0 b exp(bs ) = 1 + a a exp( bs ) 0 x < 2 b b 1 < x <0 exp(bx ) 2 = 1 1 exp( bx ) 0 x< 2 (c) The probability that 1 X 2 is

FX (2 ) FX (1) =

1 [exp( b ) exp( 2b )] 2

page8-25

Problem 8.23 Show that the expression for the variance of a random variable can be expressed in terms of

the first and second moments as

Var( X ) = E X 2 (E[ X ])

[ ]

Solution

Var( X ) = E ( X E( X ))

2

[ = E(X

[ ] = E[X ] (E[X ])

2

2 XE( X ) + (E[ X ])

2

page8-26

Problem 8.24 A random variable R is Rayleigh distributed with its probability density function given by r 2 0r< exp( r / 2b ) f R (r ) = b 0 otherwise

(a) Determine the corresponding distribution function (b) Show that the mean of R is equal to (c) What is the mean-square value of R? (d) What is the variance of R?

b / 2

FR (r ) = f R (s )ds

r 0

s2 s exp ds 2b b

s2 r = exp 2b 0 2 = 1 exp r 2b

E[R ] = sf R (s )ds

0

= =

s2 s2 exp 2b ds b

1 s2 1 2b s 2 exp 0 2b ds b 2b The bracketed expression is equivalent to the evaluation of the half of the variance of a zero-mean Gaussian random variable which we know is b in this case, so

E[R ] =

2b 1 (b ) = b b 2 2

page8-27

E R 2 = s 2 f R (s )ds

0

[ ]

s2 s3 exp 2b ds 0 b = s 2 FR (s ) 2 s FR (s ) ds 0 0 =

s2 = s 2 FR (s ) 2 s1 exp ds 2b 0 0

= s 2 (FR (s ) 1) + 2b f R (s )ds 0 0

= 2b

Var (R ) = E R 2 (E[R ])

[ ]

2 2

= 2b b 2 = b 2 2

page8-28

1 , f Z ( z ) = 2 0 , 0 z < 2 otherwise

The two random variables X and Y are related to Z by X = sin(Z) and Y = cos(Z). (a) Determine the probability density functions of X and Y. (b) Show that X and Y are uncorrelated random variables. (c) Are X and Y statistically independent? Why?

Solution (a) The distribution function of X is formally given by x 1 0 1 < x < 1 FX ( x ) = P[ 1 X x ] 1 x 1 Analogous to Example 8.8, we have P sin 1 ( x ) Z 2 + sin 1 ( x ) P( 1 X x ) = 1 1 1 + P 0 Z sin ( x ) + P sin ( x ) Z 2

1 x 0

] [

0 x 1

1 x 0 0 x 1

1 x 1

where the second line follows from the fact that the probability for a uniform random variable is proportional to the length of the interval. The distribution of Y follows from a similar argument (see Example 8.8).

page8-29

E[ XY ] = E[sin (Z ) cos(Z )] sin (2 Z ) = E 2 1 2 = sin (2 z )dz 4 0 2 1 = cos(2 z ) = 0 0 8

Thus X and Y are uncorrelated. (c) The random variables X and Y are not statistically independent since

Pr[X Y ] Pr[X ]

page8-30

Problem 8.26 A Gaussian random variable has zero mean and a standard deviation of 10 V. A constant

voltage of 5 V is added to this random variable. (a) Determine the probability that a measurement of this composite signal yields a positive value. (b) Determine the probability that the arithmetic mean of two independent measurements of this signal is positive.

Solution (a) Let Z represent the initial Gaussian random variable and Y the composite random variable. Then

Y = 5+ Z

f y (y) =

2 1 2 exp ( y ) 2 2

}

}

where corresponds to a mean of 5V and corresponds to a standard deviation of 10V. The probability that Y is positive is

P[Y > 0] = =

1 2 0 exp ( y ) 2 dy 2 2 1 exp s 2 ds 2

Making the substitutions for and , we have P[Y>0] = Q(- ). We note that in Fig. 8.11, the values of Q(x) are not shown for negative x; to obtain a numerical result, we use the fact that Q(-x) = 1- Q(x). Consequently, Q(-) = 1- 0.3 = 0.7.

page8-31

(b) Let W represent the arithmetic mean of two measurements Y1 and Y2, that is

W= Y1 + Y2 2

It follows that W is a Gaussian random variable with E[W] = E[Y] = 5. The variance of W is given by

Var (W ) = E (W E(W ) )

] ])

Y1 + Y2 E(Y1 ) + E(Y2 ) 2 = E 2 2 =

([

The first two terms correspond to the variance of Y. The third term is zero because the measurements are independent. Making these substitutions, the variance of W reduces to

Var[W ] =

page8-32

Problem 8.27

X (t ) = sin(2Wt )

in which the frequency W is a random variable with the probability density function

1 fW ( w) = B 0

Show that X(t) is nonstationary.

0<w< B otherwise

0 x < 0 FX ( 0 ) (x ) = 1 x 0

At time t = 1, X (1) = sin(2w) , and the distribution of X(1) is clearly not a step function so

FX (1) ( x ) FX (0 ) ( x )

And the process X(t) is not first-order stationary, and hence nonstationary.

page8-33

Problem 8.28

X (t ) = A cos(2f c t )

where the frequency is constant and the amplitude A is uniformly distributed:

1 f A (a ) = 0

Solution At time t = 0, X(0) = A, and FX(0)(0) is uniformly distributed over 0 to 1. At time t = (4fc)-1, X( (4fc)-1) = 0 and

1 X 4f c

(x ) = (0)

Thus, FX ( 0) ( x) FX (1 / 4 fc ) ( x) and the process X(t) is not stationary to first order. Hence not strictly stationary.

page8-34

X (t ) = A cos(2f c t )

where A is a Gaussian random variable of zero mean and variance 2. This random process is applied to an ideal integrator, producing an output Y(t) defined by

Y (t ) = X ( )d

0

(a) Determine the probability density function of the output at a particular time. (b) Determine whether or not is stationary.

Y (t ) = X ( )d

t 0

= A cos(2f c )d

t 0

A sin (2f c t ) 2f c

At time t0, it follows that Y (t 0 ) is Gaussian with zero mean, and variance

(b) No, the process Y(t) is not stationary as FY (t0 ) FY (t1 ) for all t1 and t0.

page8-35

Problem 8.30 Prove the following two properties of the autocorrelation function RX() of a random process

X(t): (a) If X(t) contains a dc component equal to A, then RX() contains a constant component equal to A2. (b) If X(t) contains a sinusoidal component, then RX() also contains a sinusoidal component of the same frequency.

Solution (a) Let Y (t ) = X (t ) A and Y(t) is a random process with zero dc component. Then

E[X (t )] = A

and

R X ( ) = E[X (t )X (t + )] = E[(( X (t ) A) + A)(( X (t + ) A) + A)] = RY ( ) + 0 + 0 + A2

And thus RX() has a constant component A2. (b) Let X (t ) = Y (t ) + A sin (2f c t ) where Y(t) does not contain a sinusoidal component of frequency fc.

R X ( ) = E[ X (t )X (t + )]

= E (Y (t ) + A sin (2f c t ))(Y (t + ) + A sin (2f c t )) + E A2 sin (2f c t ) sin (2f c (t + )) = RY ( ) + K + = RY ( ) + A [cos 2f ct + cos 2f c (2t + ) + ] 2

2

]]

A2 cos(2f c ) 2

page8-36

Yn = Yn 1 + Wn

n = , -1, 0, +1,

where the zero-mean random process Wn is stationary with autocorrelation function RW(k) = 2(k). What is the autocorrelation function Ry(k) of Yn? Is Yn a wide-sense stationary process? Justify your answer.

Solution We partially address the question of whether Yn is wide-sense stationary (WSS) first by noting that

= E[Yn1 ] + E[Wn ]

since E[Wn] = 0. To be WSS, the mean of the process must be constant and consequently, we must have that E[Yn] = 0 for all n, to satisfy the above relationship. We consider the autocorrelation of Yn in steps. First note that RY(0) is given by

and that RY(1) is

RY (1) = E[YnYn+1 ]

[ ]

= E[Yn (Yn + Wn )]

= E Yn2 + E[YnWn ]

[ ]

Although not explicitly stated in the problem, we assume that Wn is independent of Yn, thus E[YnWn] = E[Yn]E[Wn] = 0, and so

RY (1) = RY (0)

We prove the result for general positive k by assuming RY(k) = kRY(0) and then noting that

RY (k + 1) = E[YnYn+k +1 ]

page8-37

Problem 8.31 continued To evaluate this last expression, we note that, since Yn = Yn1 + Wn = 2Yn2 + Wn1 + Wn = 3Yn3 + 2Wn2 + Wn1 + Wn =L we see that Yn only depends on Wk for k n. Thus E[YnWn+k] = 0. Thus, for positive k, we have

RY (k + 1) = RY (k ) = k +1 RY (0)

Using a similar argument, a corresponding result can be shown for negative k. Combining the results, we have

RY (k ) = RY (0)

k

Since the autocorrelation only depends on the time difference k, and the process is widesense stationary.

page8-38

Problem 8.32

Find the power spectral density of the process that has the autocorrelation function

2 1 R X ( ) = 0

<1

otherwise

Solution The Wiener-Khintchine relations imply the power spectral density is given by the Fourier transform of RX(), which is (see Appendix 6)

S X ( f ) = 2 sinc 2 ( f )

page8-39

Problem 8.33. A random pulse has amplitude A and duration T but starts at an arbitrary time t0. That is,

the random process is defined as

X (t ) = Arect (t + t 0 )

where rect(t) is defined in Section 2.9. The random variable t0 is assumed to be uniformly distributed over [0,T] with density

1 f t0 ( s ) = T 0

0 s T otherwise

(a) What is the autocorrelation function of the random process X(t)? (b) What is the spectrum of the random process X(t)?

Solution First note that the process X(t) is not stationary. This may be demonstrated by computing the mean of X(t) for which we use the fact that

f X ( x) =

( x | s ) f t0 ( s )ds

xf

( x | t 0 )dx t0 t t0 + T otherwise

A = 0 Consequently, we have

E[X (t )] =

E[ X (t ) | s] f

t0

0 At / T = A(2 t / T ) 0

Thus the mean of the process is dependent on t, and the process is nonstationary.

page8-40

We take a similar approach to compute the autocorrelation function. First we break the situation into a number of cases: i) For any t < 0, s < 0, t > 2T, or s > 2T, we have that

E[X (t ) X ( s)] = 0

ii) For 0 t < s 2T, we first assume t0 is known

A2 E[X (t ) X ( s ) | t 0 ] = 0 A2 = 0 t > t0 , s < t0 + T , 0 < t0 < T otherwise max(s T ,0) < t 0 < min(t , T ) otherwise

E[ X (t ) X ( s )] = E[ X (t ) X ( s ) | w] f t0 ( w)dw

= =

1 A2 dw T max( 0 , s T )

min( t ,T )

A2 max{{min(t , T ) max(0, s T )},0} T where the second maximum takes care of the case where the lower limit on the integral is greater than the upper limit. iii) For 0 s < t 2T, we use a similar argument to obtain A2 E[X (t ) X ( s ) | t 0 ] = 0 and

E[ X (t ) X ( s )] =

page8-41

Problem 8.33 continued Combining all of these results we have the autocorrelation is given by A2 T max{{min(t , T ) max(0, s T )},0} 0 t < s 2T 2 A E[ X (t ) X ( s )] = max{{min(s, T ) max(0, t T )},0} 0 s < t 2T T otherwise 0 This result depends upon both t and s, not just t-s, as one would expect for a nonstationary process.

page8-42

Problem 8.34 Given that a stationary random process X(t) has an autocorrelation function RX() and a

power spectral density SX(f), show that: (a) The autocorrelation function of dX(t)/dt, the first derivative of X(t) is equal to the negative of the second derivative of RX(). (b) The power spectral density of dX(t)/dt is equal to 42f2SX(f). Hint: Use the results of Problem 2.24.

dX (t ) , and from the Wiener-Khintchine relations, we know the dt autocorrelation of Y(t) is the inverse Fourier transform of the power spectral density of Y. Using the results of part (b),

RY ( f ) = F 1 [SY ( f )] = F 1 4 2 f 2 S X ( f )

= F 1 ( j 2f ) 2 S X ( f )

from the differential properties of the Fourier transform, we know that differentiation in the time domain corresponds to multiplication by j2f in the frequency domain. Consequently, we conclude that

RY ( f ) = F 1 ( j 2f ) 2 S X ( f ) = d R X ( ) d 2

2

page8-43

dX (t ) , then the spectrum of Y(t) is given by (see Section 8.8) dt SY ( f ) = lim

2 1 Y E HT ( f ) T 2T

Y where H T ( f ) is the Fourier transform of Y(t) from T to +T. By the properties of Y Fourier transforms H T ( f ) = ( j 2f ) H TX ( f ) so we have

SY ( f ) = lim

2 1 Y E HT ( f ) T 2T 2 1 E ( j 2f ) H TX ( f ) = lim T 2T 2 1 4 2 f 2 E H TX ( f ) = lim T 2T 2 2 = 4 f S X ( f )

Note that the expectation occurs at a particular value of f; frequency plays the role of an index into a family of random variables.

page8-44

Problem 8.35 Consider a wide-sense stationary process X(t) having the power spectral density SX(f)

shown in Fig. 8.26. Find the autocorrelation function RX() of the process X(t).

Solution The Wiener-Khintchine relations imply the autocorrelation is given by the inverse Fourier transform of the power spectral density, thus

R( ) = S X ( f ) exp( j 2ft )df

1

= (1 f ) cos(2ft )df

0

where we have used the symmetry properties of the spectrum to obtain the second line. Integrating by parts, we obtain

1 sin (2f ) sin (2f ) R X ( ) = (1 f ) df + 2 0 0 2 1

R X ( ) = = 2 sin 2 ( ) (2 )2

1 2

sinc 2 ( )

page8-45

Problem 8.36 The power spectral density of a random process X(t) is shown in Fig. 8.27.

(a) (b) (c) (d)

Determine and sketch the autocorrelation function RX() of the X(t). What is the dc power contained in X(t)? What is the ac power contained in X(t)? What sampling rates will give uncorrelated samples of X(t)? Are the samples statistically independent?

Solution (a) Using the results of Problem 8.35, and the linear properties of the Fourier transform R( ) = 1 + 1 2 sinc 2 ( f 0 ) (b) The dc power is given by power centered on the origin

dc power = lim S X ( f )df

0

= lim ( f )df

0

=1 (c) The ac power is the total power minus the dc power ac power = RX (0) dc power

= R X ( 0) 1 =

1 2

(d) The correlation function RX() is zero if samples are spaced at multiples of 1/f0.

page8-46

Problem 8.37 Consider the two linear filters shown in cascade as in Fig. 8.28. Let X(t) be a stationary process with autocorrelation function RX(). The random process appearing at the first filter output is V(t) and that at the second filter output is Y(t). (a) Find the autocorrelation function of V(t). (b) Find the autocorrelation function of Y(t). Solution Expressing the first filtering operation in the frequency domain, we have

V ( f ) = H1 ( f ) X ( f )

where H1(f) is the Fourier transform of h1(t). From Eq. (8.87) it follows that the spectrum of V(t) is SV ( f ) = H 1 ( f ) S X ( f )

2

By analogy, we have

SY ( f ) = H 2 ( f ) SV ( f )

2

= H 2 ( f ) H1 ( f ) S X ( f )

2 2

Consequently, apply the convolution properties of the Fourier transform, we have RY ( ) = g 2 (t ) g1 (t ) * RX ( f ) where * denotes convolution; g2(t) and g1(t) are the inverse Fourier transforms of |H2(f)|2 and |H1(f)|2, respectively.

page8-47

Problem 8.38 The power spectral density of a narrowband random process X(t) is as shown in Fig. 8.29.

Find the power spectral densities of the in-phase and quadrature components of X(t), assuming fc = 5 Hz.

Solution From Section 8.11, the power spectral densities of the in-phase and quadrature components are given by

S ( f + f c ) + S ( f f c ) S N1 ( f ) = S N0 ( f ) = 0

f <B 0B

f 1 2 f S N I ( f ) = S NQ ( f ) = 2 3 2 0

otherwise

page8-48

Problem 8.39 Assume the narrow-band process X(t) described in Problem 8.38 is Gaussian with zero 2 mean and variance X .

(a) Calculate X . (b) Determine the joint probability density function of the random variables Y and Z obtained by observing the in-phase and quadrature components of X(t) at some fixed time.

2

2 X = R(0 ) = S ( f )df

1 1 = 2 b1h1 + b2 h2 2 2 1 1 = 2 2 1 + 1 1 2 2 = 3 watts

2 (b) The random variables Y and Z have zero mean, are Gaussian and have variance X . If Y and Z are independent, the joint density is given by

f Y ,Z (Y , Z ) = =

2 1 2 exp y exp z 2 2 2 X 2 X 2 X 2 X

y2 + z2 exp 2 2 2 X 2 X 1

page8-49

Problem 8.40 Find the probability that the last two digits of the cube of a natural number (1, 2, 3, )

will be 01.

Solution Let a natural number be represented by concatenation xy where y represents last two digits and x represents the other digits. For example, the number 1345 has y = 45 and x = 13. Then

where we have used the binomial expansion of (a+b)3. The last digits of the first three terms on the right are clearly 00. Consequently, it is the last two digits of y3 which determines the last two digits of (xy)3. Checking the cube of all two digit numbers for 00 to 99, we find that: (a) y3 ends in 1, only if y ends in 1; and (b) only the number (01)3 gives 01 as the last two digits. From this counting argument, the probability is 0.01.

page8-50

Problem 8.41 Consider the random experiment of selecting a number uniformly distributed over the range {1, 2, 3, , 120}. Let A, B, and C be the events that the selected number is a multiple of 3, 4, and 6, respectively. a) What is the probability of event A, i.e. P[A]? b) What is P[B]? c) What is P A B ?

d) What is P[ A B ] ? e) What is

P[ A C ] ?

P (B ) =

40 1 = 120 3

(b)

30 1 = 120 4 12 1 = 120 10

(c) (d)

P( A B ) =

P ( A B ) = P ( A) + P (B ) P ( A B ) 20 + 15 6 29 =1 +1 1 = = 3 4 10 60 60 P( A C ) = P(C ) =

(e)

20 1 = 120 6

page8-51

a) How many such messages are there? b) How many such messages are there that contain exactly four 1s? c) Suppose the 10th bit is not independent of the others but is chosen such that the modulo-2 sum of all the bits is zero. This is referred to as an even parity sequence. How many such even parity sequences are there? d) If this ten-bit even-parity sequence is transmitted over a channel that has a probability of error p for each bit. What is the probability that the received sequence contains an undetected error?

Solution (a) A message corresponds to a binary number of length 10, there are thus 210 possibilities. (b) The number of messages with four 1s is

(c) Since there are only 9 independent bits in this case, the number of such message is 29 . (d) The probability of an undetected error corresponds to the probability of 2, 4, 6, 8, or 10 errors. The received message corresponds to a Bernoulli sequence, so the corresponding error probabilities are given by the binomial distribution and is

10 10 10 10 10 2 8 6 4 2 p (1 p ) + p 4 (1 p ) + p 6 (1 p ) + p 8 (1 p ) + p10 10 8 6 4 2

page8-52

Problem 8.43 The probability that an event occurs at least once in four independent trials is equal to

0.59. What is the probability of occurrence of the event in one trial, if the probabilities are equal in all trials?

Solution The probability that the event occurs on a least one trial is 1 minus the probability that the event does not occur at all. Let p be the probability of occurrence on a single trial, so 1-p is the probability of not occurring on a single trial. Then P[at least one occurence] = 1 (1 p )

4

page8-53

Problem 8.44 The arrival times of two signals at a receiver are uniformly distributed over the interval

[0,T]. The receiver will be jammed if the time difference in the arrivals is less than . Find the probability that the receiver will be jammed.

Solution Let X and Y be random variables representing the arrival times of the two signals. The probability density functions of the random variables are

1 f X (x ) = T 0, 0< x<T otherwise

and fY(y) is similarly defined. Then the probability that the time difference between arrivals is less than is given by

P[ X Y < ] = P[ X Y < | X > Y ]P[X > Y ] + P[ X Y < | Y > X ]P[Y > X ]

= P[ X Y < | X > Y ]

where the second line follows from the symmetry between the random variables X and Y, namely, P[X > Y] = P[Y > X]. If we only consider the case X > Y, then we have the conditions: 0 < X < T and 0 < Y < X < +Y. Combining these conditions we have Y < X < min(T, +Y). Consequently,

P[ X Y < ] =

T 0

min (T , + y )

f

y

( x) f Y ( y )dx dy

=

0

min (T , + y )

1 dx dy T

1 T2

{min(T , + y ) y}dy

0

T T

1 y2 Ty , y = 2 min 2 T 0 1 = min , 2 T

page8-54

Problem 8.45 A telegraph system (an early version of digital communications) transmits either a dot or

dash signal. Assume the transmission properties are such that 2/5 of the dots and 1/3 of the dashes are received incorrectly. Suppose the ratio of transmitted dots to transmitted dashes is 5 to 3. What is the probability that a received signal as the transmitted if: a) The received signal is a dot? b) The received signal is a dash?

Solution (a) Let X represent the transmitted signal and Y represent the received signal. Then by application of Bayes rule

P(Y = dot ) = P( X = dot | No error )P( No dot error) + P( X = dash | error )P(dash error ) =5 3 + 3 1 8 5 8 3 =3 +1 =1 8 8 2

( ) ( )( )

(b) Similarly, P[Y = dash ] = P[ X = dash | no error]P(no dash error ) + P( X = dot )P[dot error] = 3 2 + 5 2 8 3 8 5 =2 +2 = 1 8 8 2

[ ]

page8-55

Problem 8.46 Four radio signals are emitted successively. The probability of reception for each of them is independent of the reception of the others and equal, respectively, 0.1, 0.2, 0.3 and 0.4. Find the probability that k signals will be received where k = 1, 2, 3, 4. Solution For one successful reception, the probability is given by the sum of the probabilities of the four mutually exclusive cases

P = p1 (1 p2 )(1 p3 )(1 p4 ) +

= .1 .8 .7 .6 + .9 .2 .7 .6 + .9 .8 .3 .6 + .9 .8 .7 .4 = 0.4404

P = p1 p2 (1 p3 )(1 p4 ) + p1 (1 p2 ) p3 (1 p4 ) + p1 (1 p2 )(1 p3 ) p4 +

(1 p1 ) p2 p3 (1 p4 ) + (1 p1 ) p2 (1 p3 ) p4 + (1 p1 )(1 p2 ) p3 p4

P = p1 p2 p3 (1 p4 ) + p1 (1 p2 ) p3 p4 + p1 p2 (1 p3 ) p4 + (1 p1 ) p2 p3 p4 = .1 .2 .3 .6 + .1 .8 .3 .4 + .1 .2 .7 .4 + .9 .2 .3 .4 = 0.0404

P = p1 p2 p3 p4 = .1 .2 .3 .4 = 0.0024

page8-56

Problem 8.47 In a computer-communication network, the arrival time between messages is modeled

with an exponential distribution function, having the density

1 e f T ( ) = 0

0

otherwise

a) What is the mean time between messages with this distribution? b) What is the variance in this time between messages?

Solution (Typo in problem statement, should read fT()=(1/)exp(-/) for >0) (a) The mean time between messages is

E[T ] = f T ( )d

0

exp( / )d 0

= exp( / ) 0 + exp( / )d

0

= 0 exp( / ) 0 =

where the third line follows by integration by parts. (b) To compute the variance, we first determine the second moment of T

ET

[ ] =

2 0

f T ( )d

2 exp( / )d 0

0

0

page8-57

The variance is then given by the difference of the second moment and the first moment squared (see Problem 8.23)

Var (T ) = E T 2 (E[T ]) = 22 2 = 2

[ ]

page8-58

Problem 8.48 If X has a density fX(x), find the density of Y where a) Y = aX + b for constants a and b. 2 b) Y = X . c) Y = X , assuming X is a non-negative random variable. Solution (a) If Y = aX + b , using the results of Section 8.3 for Y = g(X)

dg 1 ( y ) f Y ( y ) = f X (g ( y ) ) dy

1

y b 1 = fX a a (b) If Y = X 2 , then

fY ( y ) = f X y + f X +

( (

1 y 2 y

))

(c) If Y = X , then we must assume X is positive valued so, this is a one-to-one mapping and

fY ( y ) = f X y 2 2 y

( )

page8-59

Problem 8.49 Let X and Y be independent random variables with densities fX(x) and fY(y), respectively. Show that the random variable Z = X+Y has a density given by

f Z ( z) =

Hint:

( z s ) f X ( s )ds

P[Z z ] = P[ X z, Y z X ]

FZ ( z ) =

z zx

( x) f Y ( y )dydx

To differentiate this result with respect to z, we use the fact that if g ( z ) = h( x, z )dx

a b

b

(1)

zx

h ( x, z ) =

( x) f Y ( y )dy

f Z ( z) = d FZ ( z ) dz z ( ) z zz d z x dz f X ( x) f Y ( y )dy dx + f X ( z ) f Y ( y )dy f X () f Y ( y )dy 0 = dz dz = d dz

z z x

( x) f Y ( y )dy dx

page8-60

where the second term of the second line is zero since the random variables are positive, and the third term is zero due to the factor zero. Applying the differentiation rule a second time, we obtain

f Z ( z) = =

0 + f f

X

( x) f Y ( z x)

d ( z x) d () dx f X ( x) f Y () dz dz

( x) f Y ( z x)dx

P[Z z | X = x] = P[X + Y z | X = x]

where the third equality follows from the independence of X and Y. By differentiating both sides with respect to z, we see that f Z | X ( z | x) = f Y ( z x) By the properties of conditional densities f Z , X ( z , x) = f X ( x) f Z | X ( z | x) = f X ( x) f Y ( z x) Integrating to form the marginal distribution, we have

f Z ( z) =

( x) f Y ( z x)dx

If Y is a positive random variable then fY(z-x) is zero for x > z and the desired result follows.

page8-61

Z (t ) = X (t )Y (t )

where X(t) and Y(t) are independent zero-mean random processes with

RX ( ) = a1e

and

RY ( ) = a2 e

RZ ( ) = E[Z (t )Z (t + )]

= E[ X (t )X (t + )Y (t )Y (t + )] = E[ X (t ) X (t + )]E[Y (t )Y (t + )] = RX ( )RY ( )

By the Wiener-Khintchine relations, the spectrum of Z(t) is given by S Z ( f ) = F 1 [R X ( )RY ( )] = F 1 [a1a2 exp( (1 + 2 ) = 2a1a2 (1 + 2 ) 2 (1 + 2 ) 2 + (2f )

)]

where the last line follows from the Fourier transform of the double-sided exponential (See Example 2.3).

page8-62

X (t ) = sin (2f c t )

where the frequency fc is a random variable uniformly distributed over the interval [0,W]. Show that X(t) is nonstationary. Hint: Examine specific sample functions of the random process X(t) for, say, the frequencies W/4, W/2, and W.

Solution To be stationary to first order implies that the mean value of the process X(t) must be constant and independent of t. In this case, E[ X (t )] = E[sin (2f c t )]

W

cos(2wt ) = 2Wt 0 =

1 cos(2Wt ) 2Wt

This mean value clearly depends on t, and thus the process X(t) is nonstationary.

page8-63

Problem 8.52 The oscillators used in communication systems are not ideal but often suffer from a distortion known as phase noise. Such an oscillator may be modeled by the random process

Y (t ) = A cos(2f c t + (t ) )

where

(t )

is a slowly varying random process. Describe and justify the conditions on the random process

(t )

Solution The first condition for wide-sense stationary process is a constant mean. Consider t = t0, then

E[Y (t 0 )] = E[ A cos(2f c t0 + (t0 ) )]

In general, the function cos takes from values -1 to +1 when varies from 0 to 2. In this case corresponds to 2fct0 + (t0). If (t0) varies only by a small amount then will be biased toward the point 2fct0 + E[(t0)], and the mean value of E[Y(t0)] will depend upon the choice of t0. However, if (t0) is uniformly distributed over [0, 2] then 2fct0 + (t0) will be uniformly distributed over [0, 2] when considered modulo 2, and the mean E[Y(t0)] will be zero and will not depend upon t0. Thus the first requirement is that (t) must be uniformly distributed over [0,2] for all t. The second condition for a wide-sense stationary Y(t) is that the autocorrelation depends only upon the time difference

E[Y (t1 )Y (t 2 )] = E[A cos(2f c t1 + (t1 ) )A cos(2f c t 2 + (t 2 ) )] A2 E[cos(2f c (t1 + t 2 ) + (t1 ) + (t 2 ) ) + cos(2f c (t1 t 2 ) + (t1 ) (t 2 ) )] = 2

where we have used the relation cos A cos B = 1 2 (cos( A + B) + cos( A B) ) . In general, this correlation does not depend solely on the time difference t2-t1. However, if we assume: We first note that if (t1) and (t2) are both uniformly distributed over [0,2] then so is = (t1 ) + (t 2 ) (modulo 2), and

E[cos(2f c (t1 + t 2 ) + )] =

1 2

2

cos(2f

0

(t1 + t 2 ) + )d

(1)

=0

page8-64

We consider next the term RY(t1,t2)= E[cos(2f c (t1 t 2 ) + (t1 ) (t 2 ) )] and three special cases: (a) if t = t1-t2 is small then (t1 ) (t 2 ) since (t) is a slowly varying process, and

A2 cos(2f c (t1 t 2 ) ) 2 (b) if t is large then (t1) and (t2) should be approximately independent and (t1 ) (t 2 ) would be approximately uniformly distributed over [0,2]. In this case RY (t1 , t 2 ) 0 using the argument of Eq. (1). RY (t1 , t 2 ) =

(t1 ) (t 2 ) g (t1 t 2 )

for some arbitrary function g(t). Under these conditions the random process Y(t) will be wide-sense stationary.

page8-65

X (t ) = A sin (0.3t ) + N (t )

where N(t) is a stationary Gaussian process of zero mean and variance 2. (a) What are the density functions of the random variables X1 and X2 where

X 1 = X (t ) t =1 X 2 = X (t ) t =2

(b) The noise process N(t) has an autocorrelation function given by

RN ( ) = 2 exp(

What is the joint density function of X1 and X2, that is,

f X1 , X 2 ( x1 , x2 ) ?

E[ X (t1 )] = E[ A sin (0.3 ) + N (t1 )]

= A sin (0.3 ) = A sin(0.3 ) + E[N (t1 )]

Since X1 is equal to N(t1) plus a constant, the variance of X1 is the same as that of N(t1). In addition, since N(t1) is a Gaussian random variable, X1 is also Gaussian with a density given by 1 f X1 ( x ) = exp{ ( x 1 ) / 2 2 } 2 where 1 = E[ X (t1 )] . By a similar argument, the density function of X2 is

f X 2 ( x) = 1 exp ( x 2 ) / 2 2 2

where 2 = A sin(0.6 ) .

page8-66

(b) First note that since the mean of X(t) is not constant, X(t) is not a stationary random process. However, X(t) is still a Gaussian random process, so the joint distribution of N Gaussian random variables may be written as Eq. (8.90). For the case of N = 2, this equation reduces to

f X (x) = 1 2

1/ 2

exp (x )1 (x )T / 2

where is the 2x2 covariance matrix. Recall that cov(X1,X2) =E[(X1-1)(X2-2)], so that

cov( X 1 , X 1 ) cov( X 1 , X 2 ) = cov( X 2 , X 1 ) cov( X 2 , X 2 ) R (0) RN (1) = N RN (1) RN (0) 2 2 exp(1) = 2 2 exp( 1)

= 4 (1 2 )

and

1 =

1 1 2 (1 )

2

Making these substitutions into the above expression, we obtain upon simplification

f X1 , X 2 ( x1 , x2 ) = 1 2 2 ( x 1 ) 2 + ( x2 2 ) 2 2 ( x1 1 )( x2 2 ) exp 1 2 2 (1 2 ) 1 2

page8-67

- Normal Distribution - Wikipedia, The Free EncyclopediaUploaded byLieu Dinh Phung
- 320Uploaded bySridhar Miriyala
- 2-Way-TOS-Reading and Writing -.xlsxUploaded byAngelica Babsa-ay Asiong
- Statistics Problems 2 Chapter 9Uploaded byALyzha Mae Lacuesta
- Time History Ground Motion Scaling in EtabsUploaded byRaleigh Gesulga Labarrete
- x Ray Diffraction and Nmr Data for the Study of the Location of Idebenone and Idebenol in Model MembranesUploaded byJoão Honorato
- An Analysis of Pressure Fluctuations in a CFB of Heavy MineralsUploaded byDr-Samy El-Behery
- Normal DistributionUploaded byInnocentOvick
- hw4Uploaded byjamesyu
- 100524Uploaded byvol2no5
- Shecdiling.pdfUploaded byshijinrajagopal
- 2.2Uploaded bySandeepVerma
- The Measurement of Cholinesterase Activities as a BiomarkerUploaded bymiriade65
- Project 2010Uploaded byrameshpani
- rr311801-probability-and-statisticsUploaded bySRINIVASA RAO GANTA
- qch11-12-13Uploaded bylightknow
- 25101-720_UE Radio Transmission and Reception (FDD)Uploaded byJason Liew
- Biological_Variation_update_ed.pptUploaded bymaimaimimai
- 2013_Ch.13_Notes-2Uploaded byFaith
- Chapter 6Uploaded byjernejaj
- Introductory Statistics NotesUploaded bystefaniasiatra
- 04074713Uploaded byDragonsito
- 2007-01Uploaded byLuminita Duta
- Earnings Quality MetricsUploaded byThe Actuary
- spectrum estimation.pptUploaded byDr Mohammed
- PowerPoint Slides to PCS Chapter 01 Part BUploaded bySudhanshu Verma
- Data Tables & Monte Carlo Simulations in Excel – a Comprehensive GuideUploaded byhomomysticus
- Wind Loads on BridgesUploaded byaerodanger
- Sampling n Sampling DistributionUploaded bystu_aro
- Determination of CharacteristicUploaded byMahesh Kanike

- (LAB)Uploaded byAbd El-Hai Mohammed
- V2N1-001Uploaded byAbd El-Hai Mohammed
- [SERSC] Author GuidelinesUploaded byAbbas Abbasi
- SocEncounter2013(LAB)Uploaded byAbd El-Hai Mohammed
- joules_ugUploaded byAbd El-Hai Mohammed
- Request for SupervisorUploaded byAbd El-Hai Mohammed
- Questions TipsUploaded byAbd El-Hai Mohammed
- Spring 2018 UG, 17 Feb 2018Uploaded byAbd El-Hai Mohammed
- New Text Document (2)نن.txtUploaded byAbd El-Hai Mohammed
- Solution EPE 121 Assignment Prpblems ChapterUploaded byAbd El-Hai Mohammed
- Joules Cmd RefUploaded byAbd El-Hai Mohammed
- Solution Chapter 2Uploaded byAbd El-Hai Mohammed
- Solutiuon EPE121 Electrical MachinesUploaded byAbd El-Hai Mohammed
- 0623163300Uploaded byAbd El-Hai Mohammed
- بسم الله الرحمن الرحيمUploaded byAbd El-Hai Mohammed
- Chapter 7 sUploaded byAbd El-Hai Mohammed
- yourpreziUploaded byAbd El-Hai Mohammed
- wordsUploaded byAbd El-Hai Mohammed
- wordsUploaded byAbd El-Hai Mohammed
- wordsUploaded byAbd El-Hai Mohammed
- wordsUploaded byAbd El-Hai Mohammed
- New Text DocumentUploaded byAbd El-Hai Mohammed
- dont download it cause copy rightUploaded byAbd El-Hai Mohammed
- 0301.pdfUploaded byAbd El-Hai Mohammed
- 1Uploaded byAbd El-Hai Mohammed

- SIBL Final Copy CopyUploaded byAsiburRahman
- cpphtp4_PPT_18Uploaded byMahesh Kumar
- Seal FluentUploaded byrhbazzi
- Ericsson Consono MD110 PBX Managemen Module GuideUploaded bydgiak73
- L. Prasad, IIM, BangaloreUploaded bydeepasridhar
- Chapter1SpinelloUploaded byShendelzare Silkwood
- CH21Uploaded byJesha Librea
- scn 400 - final project lesson planUploaded byapi-282343909
- Transformation (Genetics) - Wikipedia, The Free EncycloTransformation (genetics)pediaUploaded byEmad Manni
- The Psycho Pathology of Walrasian Marxism-Alan FreemanUploaded bychavismo
- ku honors program resumeUploaded byapi-439608574
- Psychiatric NursingUploaded byJohn Christopher Celestino
- Design of Machine ElementsUploaded byNeerav Saxena
- Foucalt and NeoliberalismUploaded byJorge Giannareas
- atomic theory modelsUploaded byapi-318918663
- rethoric essayUploaded byapi-302486962
- depm 622 a3- final draftUploaded byapi-237484773
- City-Building-Scenery-Real-Estate-PowerPoint-Template-1.pptxUploaded byAhmad Danil
- Machine DreamsUploaded byNick Serpe
- gl_seismic_dp_20140313Uploaded byNirwansyah Eka Bimatara
- The Education System in the USAUploaded byNastja Vogrinc
- Bu BubububuUploaded byKen Ken Ü
- Chapter 1Uploaded bySyeda Dilawaiz
- Tutorial 1Uploaded byAmir Chip
- R3 Profile ParametersUploaded byChawalit Kittaweepwattana
- Fundamentals of Engine ControlUploaded bym21m29
- SETEC Candidature - Fondation Louis Vuitton _1Uploaded bySneha Susan John
- LPs.8th gr English.1.10-1.14Uploaded byChenelle Boatswain
- As 2134.3-1994 Recommended Practice for Chemical Analysis by Atomic Absorption Spectrometry Vapour GenerationUploaded bySAI Global - APAC
- completed essential 3 e-folio nursing 402Uploaded byapi-374818694