When the ratio of variations in the related variables is constant it is called positive correlation?

Descriptive Statistics II: Bivariate and Multivariate Statistics

Andrew P. King, Robert J. Eckersley, in Statistics for Biomedical Engineers and Scientists, 2019

2.3.1 Covariance

In Section 1.5.1 the concept of the standard deviation was introduced as a measure of variation for a single continuous or discrete variable. This concept can be extended to the case where we have multiple variables. The corresponding measure of variation for multiple variables (i.e. multivariate data) is known as covariance. Covariance is a measure of how much the variations of two variables are related. A positive covariance between two variables reveals that the paired values of both variables tend to increase together. A negative covariance reveals that there is an inverse relationship between the variables, that is, as one increases, the other tends to decrease. A zero covariance indicates that there is no link between the values of the two variables.

For two paired variables xi and yi, i=1,…,n (where n is the sample size), the covariance is defined by

(2.1)cov(x,y)=1n−1∑i=1n(x i−x¯)(yi−y ¯).

Interpretation of the magnitude of the covariance is not straightforward. A quick examination of Eq. (2.1) reveals that the covariance will be larger if the values of the variables themselves are larger. So without some form of normalization it is not possible to use covariance to determine how strong the relationship between the variables is. Pearson's correlation coefficient (see Section 2.4.1) does exactly this by normalizing the covariance by the product of the standard deviations of the variables.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780081029398000116

Patterns of residual covariance structure

Xian Liu, in Methods and Applications of Longitudinal Data Analysis, 2016

5.1.1 Compound symmetry (CS)

In Chapter 3, I described the covariance structure for the random intercept model. Specifically, if b0i is considered to be the only random effect in a linear mixed model, variance of the between-subjects random effects is constant throughout all time points and so is covariance between two subject-specific observations. Such a covariance structure is referred to as compound symmetry (CS). In longitudinal data analysis, the CS covariance structure can also be specified for the R-sided matrix.

Let σ2 be the constant variance in mean μ across repeated measurements of the response and σ1 be the constant and positive covariance between any two successive observations. Given four equally spaced time points, the residual CS covariance structure can be written as

(5.1)R=σ 2σ1σ1σ1σ1σ2σ1σ1σ1σ1σ2σ 1σ1σ1σ1σ2,

where R, as defined previously, is the residual variance–covariance matrix for within-subject repeated measurements.

Given the intimate association between covariance and correlation, Equation (5.1) can be expressed in terms of a constant correlation coefficient by factoring σ2 out, written as

(5.2)R=σ21ρρρρ1ρρ ρρ1ρρρρ1.

The variance–covariance structure of CS is the simplest and the most parsimonious pattern model designed for longitudinal analysis. This model includes only two variance parameters, σ2 and σ1. If longitudinal data follow the pattern of constant variance and constant covariance across all time points, intraindividual correlation is addressed by the specification of σ1. Given this type of data, a linear regression model including a residual variance–covariance matrix of CS can yield statistically efficient and consistent parameter estimates. In some clinical experimental designs, where repeated measurements are narrowly spaced with equal distance, the use of CS in longitudinal analysis is effective. In most occasions, however, the hypothesis of CS on residuals is too strong to specify the R-sided covariance matrix in linear regression models. In the vast majority of longitudinal data, particularly those from observational surveys, intraindividual correlation is not constant over time without the specification of the between-subjects random effects. As a result, the CS covariance structure is empirically much less applied than the other more complex pattern models.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780128013427000058

Nonlinear optimization applied to the portfolio theory

Giovanni Romeo, in Elements of Numerical Mathematical Economics with Excel, 2020

The case of two assets

We will denote the portfolio return with rp as follows (sometimes it is also referred to as the expected return of the portfolio):

rp=w1r1+w2r2.

The return is always associated to the risk of a portfolio, and finance takes from statistics its measure of risk, which is a measure of dispersion around a central tendency of the given random variable. The standard deviation or equivalently its square, the variance, is taken therefore as one of the principal risk indicators of a portfolio. See also Section 13.3.

Theory of finance introduced as well some other “combined” measures, like the portfolio returns adjusted by the risk, and within this class of risk-adjusted performance indicators, the Sharpe Ratio plays a key role.

The variance of a portfolio of two assets is calculated as follows (from the property of the variance of two variables in statistics and probability theory):

σp2=w12σ12+w22σ22+2w1w2ρ12σ1σ2

where

σ1=standarddeviation ofasset1

σ1=sta ndarddeviationofasset2

ρ12=Cov(r1,r2)σ1σ2correlationcoefficientbetwee nthetwofinancialassets.

The index of covariance Cov(r1, r2) is an index to study the relationship between two assets returns (normally the fund's returns and the benchmark's returns) which represent the statistical (or random) variables we are studying. The covariance ranges from negative values to positive values. A positive covariance indicates that the two variables tend to move together and with the same sign, a negative covariance indicates that the two variables tend to move in the opposite direction. A covariance close to zero indicates that the relation between the fund's returns and the benchmark's returns is neutral and therefore there is little relationship between the two variables.

The formula of the covariance considers the two variables (r1, r2) at the same time, and it is as follows:

Cov(r1,r2) =1N∑i=1N(r1i−μ1)⋅(r2i−μ2)

with μ1 and μ2 the expected (average) returns of the first and the second assets, respectively.

To measure the correlation between two variables, we calculate the correlation coefficient ρ12 with:

−1≤ρ12≤ +1.

Since via the covariance we can only recognize the magnitude of the relationship in absolute terms, this index helps us to investigate more about the magnitude of the relationship in relative terms. This means that ρ12 ranges from −1 to +1, and therefore we can recognize immediately the strength and the quality of the relationship between the two variables that we are studying.

When we find values of ρ12 = 1 we have got perfect positive correlation, when we find values of ρ12 = −1 we have got perfect negative correlation. Values close to zero indicate neutral correlation between the two variables, in other words they indicate independence between the two variables.

Let us represent the variances and covariances of the portfolio in the following symmetric matrix, called the matrix of variances and covariances associated to the portfolio p:

Σ=[Cov(r1,r1)Cov(r1, r2)Cov(r2,r1)Cov(r2,r2)]

then σp 2 can be represented as a quadratic form notation (see Eq. 3.4-2):

σp2=wTΣ w

which is, recalling that Cov(ri,ri)=σi2:

σp2=[w1w2] [σ12Cov(r1,r 2)Cov(r2,r1)σ22][w1w2]

σp2=[w1σ12+w2Cov(r2,r1)w1 Cov(r1,r2)+w2σ 22][w1w2]

which corresponds to a scalar, namely the sought portfolio variance:

σp2=w12σ12+w1 w2Cov(r2,r1)+w1 w2Cov(r1,r2)+w2 2σ22

=w12σ12+w22 σ22+2w1w2Cov(r1,r2)

=w12σ12+w22 σ22+2w1w2ρ12σ1σ2 .

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780128176481000086

Discrete Random Variables

Sheldon M. Ross, in Introductory Statistics (Fourth Edition), 2017

5.5.1 Covariance and Correlation

Let X and Y be random variables with, respective, expectations μx=E[X] and μy=E[Y]. The covariance of X and Y is defined by

Cov(X,Y) =E[(X−μx)(Y−μy)].

That is, the covariance of X and Y is the expected value of the product of X minus its expected value and Y minus its expected value. By expanding the product (X−μx )(Y−μy) and then using that the expected value of a sum is equal to the sum of the expected values, we can show that the covariance of X and Y can, alternatively, be expressed as the expected value of the product of X and Y minus the product of their expectations. We now give that derivation.

(5.5)Cov(X,Y)=E[(X−μx)(Y−μy)]

(5.6)=E[XY−Xμy−μ xY+μxμy]=E[XY]−E[Xμy]−E[ μxY]+E[μxμy]=E[XY]−μyE[X]−μxE[Y]+μxμy =E[XY]−μyμ x−μxμy+μxμy=E[XY]−μxμy

where the preceding used that the expected value of a constant times a random variable is just the constant times the expected value of the random variable, and that the expected value of a constant is just the constant.

To understand the significance of the covariance, consider a situation where X and Y have the property that when X is large then Y also tends to be large, and when X is small then Y also tends to be small. However, to understand the preceding, we have to first decide what it means to say that “X is large” or that “X is small.” Now, because E[X] is the average value of X, one reasonable interpretation of the preceding is that X is large when X is larger than E[X], and that X is small when X is smaller than E[X]. Thus, to say that “when X is large then Y also tends to be large” is to say that when X>E[X] then Y tends to be larger than E[Y]; and to say that when X is small then Y also tends to be small, is to say that when X<E[X] then Y tends to be smaller than E[Y]. But this is equivalent to saying that X−E[X] and Y−E[Y] typically have the same sign (that is, typically either both are positive or both are negative). But if X−E [X] and Y−E[Y] have the same sign, then their product (X−E[X ])(Y−E[Y]) will be positive, and thus its expected value E[(X−E[X])(Y−E[Y])] will be positive. That is, Cov(X,Y)>0 when X and Y tend to either both be large or both be small.

Now suppose X and Y have the property that when one is large the other tends to be small. Thus, when X> E[X] then, typically, Y<E[Y]. But then X−E[X] and Y−E[Y] typically have opposite signs (when one is positive the other is negative) and so their product will tend to be negative, and thus, E[(X−E [X])(Y−E[Y])] will be negative. Consequently, Cov(X,Y)<0 if when one is large the other tends to be small.

Definition

Say that X and Y are positively correlated if Cov(X,Y)>0 and say that they are negatively correlated if Cov(X,Y)<0.

Example 5.20

Suppose that X and Y are Bernoulli random variables, with respective parameters px and py. That is,

px=P(X=1)=1− P(X=0)py=P(Y=1 )=1−P(Y=0)

Because XY=1 if both X and Y equal 1 and XY =0 otherwise, it follows that

E[XY]=1⋅P (XY=1)+0⋅P(XY=0)=P(XY=1)=P(X=1,Y=1)

Now, it follows from Example 5.6 that E[X]=px and E[Y]=py. Consequently, we obtain, upon using (5.6), that

Cov(X, Y)=P(X=1,Y=1)−pxp y.

Now,

P(X=1,Y=1)=P(X=1)P(Y=1|X=1) =pxP(Y=1|X=1)

and so

Cov(X,Y)=pxP(Y=1|X=1)−pxpy=px( P(Y=1|X=1)−py).

Thus, in this case, Cov(X,Y)>0 is equivalent to P(Y=1|X=1)−py>0; that is, it is equivalent to P(Y=1|X=1)>P(Y=1). Thus, X and Y have a positive covariance if knowledge that X=1 (the largest possible value of X), makes it more likely that Y=1 (the largest possible value of Y). Similarly, Cov(X,Y)<0 can be shown to be equivalent to P(Y =1|X=1)<P(Y=1).

Whereas a positive value of the covariance between X and Y indicates that when one is larger than its mean then the other also tends to be larger than its mean, and a negative value indicates that when one is larger than its mean then the other tends to be smaller than its mean, the correlation measures the strength of the relationship. Its value is always between −1 and 1, with a large positive value of the correlation indicating that large values of one of the variables strongly go along with large values of the other, and a large negative value indicating that large values of one are almost always associated with small values of the other. A correlation near 0 indicates that the relationship is weak. The correlation is obtained by dividing the covariance by the product of the standard deviations of X and of Y. That is,

Definition

The correlation between X and Y, denoted as Corr(X,Y), is defined by

Corr(X,Y)=Cov(X,Y)Var(X)Var(Y)

Example 5.21

An instructor of a one week course presenting introductory material on the mathematics needed to take a probability course gives quizzes on the first and last day of class, with possible quiz grades on each being 1, 2 and 3, with 1 being the worst and 3 the best score. Based on the instructor's experience, if X is a randomly chosen student's score on the first exam and Y is his or her score on the second exam then, with p(i,j)=P(X=i,Y=j),

p(1,1)=.09,p(1,2)=.15,p(1,3)=.06p(2,1)=.10,p(2,2)=.25,p(2,3)=.15p(3,1)=.02,p(3,2)=.08,p(3,3)=.10

Thus, for instance, the probability that a student scores 2 on the first exam and 3 on the second is p(2,3)=.08 . (As a check on these 16 probabilities, note that they sum, as they must, to 1.)

Using that P(X=i)=p(i,1)+ p(i,2)+p(i,3), we see from the preceding that

P(X=1)=.30,P(X=2)=.50,P( X=3)=.20

Similarly, using that P(Y=j)=p(1,j )+p(2,j)+p(3,j) gives that

P(Y=1)=.21,P(Y=2)=.48,P(Y=3)=.31

It follows from the preceding that

E[X]=1( .30)+2(.50)+3(.20)=1.90

and, similarly,

E [Y]=.21+.96+.93=2.10

The probability mass function of XY is

P(XY=1)=p(1,1)=.09P(XY=2)=p(1,2)+p(2,1)=.25P(XY=3)=p(1,3)+p(3,1)=.08P(XY=4)=p(2,2)=.25P(XY=6)=p(2,3)+p(3,2) =.23P(XY=9)=p(3,3)=.10

Hence,

E[XY]=1(.09)+2( .25)+3(.08)+4(.25)+6(.23 )+9(.10)=4.11

Therefore,

Cov(X,Y)=E[XY]−E[X]E[Y]=4.11−1.90(2.10 )=.12

which shows a positive correlation between the two test scores. To determine the strength of this positive relationship, we compute the correlation between X and Y. To do so, first we will compute Var(X) and Var(Y). Because X2 takes on possible values 12,2 2, and 32 with, respective, probabilities .3, .5 and .2 we see that

E[X2]=1(.3)+4(.5)+9(.2)=4.1

and, similarly,

E[Y2]=.21+4(.48)+9(.31)=4.92.

Hence, upon using that the variance of any random variable is the expected value of its square minus the square of its expected value, we see that

Var(X)=4.1−(1.90)2=.49,Var(Y)=4.92−(2.10)2=.51

Thus,

Corr(X,Y)=Cov(X,Y)Var(X)Var(Y)=.12.49(.51)=.240.

Example 5.22

Suppose that X and Y are both Bernoulli (that is, their possible values are 0 and 1) random variables with the same parameter p, meaning that

P(X=1)=P(Y=1)=pP(X=0)=P(Y= 0)=1−p.

Now, from Examples 5.6 and 5.12, we have

E[X]=E[Y]=pVar(X)=Var(Y)=p(1−p)

and, from Example 5.20,

Cov(X,Y)=p(P(Y=1|X=1)− p).

Hence,

(5.7)Corr(X,Y)=Cov(X,Y)Var(X)Var(Y)

(5.8)=P(Y=1|X=1)−p1−p

where the preceding used that Var(X)Var(Y)=p(1−p)p(1−p)=p(1−p).

For instance, suppose that balls are to be sequentially selected, without replacement, from an urn containing 3 red and 5 blue balls. Let X be the indicator of the event that the first ball selected is red (that is, X=1 if the first ball is red and X=0 otherwise), and let Y be the indicator of the event that the second ball selected is red. Because, for k=1 or k=2, each of the 8 balls is equally likely to be the kth ball to be selected, it follows that

P(X=1)=P (Y=1)=3/8.

Now, given that the first ball selected is red, the second ball is equally likely to be any of the remaining 7 balls, of which 2 are red. Hence,

P(Y=1|X=1)=2/7

yielding, from (5.8), that

Corr(X,Y)=2/7−3/85/8=−5/ 565/8−1/7.

Read full chapter

URL: https://www.sciencedirect.com/science/article/pii/B9780128043172000059

On homogenization of space–time dependent and degenerate random flows

Rémi Rhodes, in Stochastic Processes and their Applications, 2007

We will outline now the main ideas of the proof. Our goal is to show that the rescaled process

εX t/ε2ω=ε∫0t/ε2b(r,Xrω,ω)ds+ε∫0t/ε2σ(r,X rω,ω)dBs

converges in law to a Brownian motion with a certain positive covariance matrix. The general strategy (see [8]) consists in finding an approximation of the first term on the right-hand side by a family of martingales and then in applying the central limit theorem for martingales. To find such an approximation, we look at the environment as seen from the particle

Yt= τt,Xtωω,

where {τt,x} is a group of measure preserving transformations on a random medium Ω (see Definition 2.1). Thanks to the particular choice of the drift, an explicit invariant measure can be found for this Markov process. The ergodicity is ensured by the geometry of the diffusion coefficient σ (see Assumptions 2.3 and 2.4). The approximation that we want to find leads us to study the equation (λ>0)

(4)λuλ−(L+Dt)uλ=b

where L+Dt coincides with the generator of the process Y on a certain class of functions (the term Dt is due to the time evolution and L is an unbounded operator on the medium Ω associated with (1)). Here there arise difficulties resulting from the time dependence. Due to the term Dt, the Dirichlet form associated with L+Dt does not satisfy any sector condition (even a weak one). However, for a suitable function b, (4) can be solved with the help of an approximating sequence of Dirichlet forms with weak sector condition. Then, usual techniques used in the static case fall short of establishing the so-called sublinear growth of the correctors . To get round this difficulty, regularizing properties of the heat kernel are used in [2,10] or [11]. Here the degeneracies of the diffusion coefficient prevent us from using such arguments. The strategy here consists in separating the time and spatial evolutions (see Assumption 2.3). We introduce a new operator whose coefficients do not depend on time. Then the spectral calculus linked to the normal operator S˜+Dt will be determined to establish the desired estimates for the solution of the equation

λvλ−(S˜+Dt)vλ=b .

Finally, with perturbation methods, we show that these estimates remain valid for the correctors .

Read full article

URL: https://www.sciencedirect.com/science/article/pii/S0304414907000178

Linear Correlation: If the amount of change in one variable tends to bear constant ratio to the amount of change in other variable then the relation is called linear correlation.

What variation has a ratio of two variables that is constant?

Definition: Direct Variation or Direct Proportion Two variables are said to be in direct variation, or direct proportion, if their ratio is constant. This type of relationship is often written as 𝑦 ∝ 𝑥 , which is read as 𝑦 is directly proportional to 𝑥 .

What is difference between positive and negative correlation?

A positive correlation exists when two variables operate in unison so that when one variable rises or falls, the other does the same. A negative correlation is when two variables move opposite one another so that when one variable rises, the other falls.

When two variables have a positive correlation?

A positive correlation is a relationship between two variables that tend to move in the same direction. A positive correlation exists when one variable tends to decrease as the other variable decreases, or one variable tends to increase when the other increases.