used to measure the size of vector; vector distance
Properties
for any n-vectors x and y, and any scalar β
homogeneity: ∥βx∥=∣β∣∥x∥
triangle inequality: ∥x+y∥≤∥x∥+∥y∥
nonnegativity: ∥x∥≥0
definiteness: ∥x∥=0only ifx=0
RMS value
Mean-square value of n-vector x is
nx12+⋯+xn2=n∥x∥2
Root-mean-square value (RMS value) is
rms(x)=nx12+⋯+xn2=n∥x∥
rms(x) gives typical value of ∣xi∣
e.g., rms(1)=1(independent ofn)
RMS value useful for comparing sizes of vectors of different lengths
Norm of block vectors
suppose a,b,c are vectors
∥(a,b,c)∥2=aTa+bTb+cTc=∥a∥2+∥b∥2+∥c∥2
so we have
∥(a,b,c)∥=∥a∥2+∥b∥2+∥c∥2=∥(∥a∥2,∥b∥2,∥c∥2)∥
Chebyshev inequality
suppose that k of the numbers ∣x1∣,…,∣xn∣ are ≥a
then k of the numbers x12,…,xn2 are ≥a2
so ∥x∥2=x12+⋯+xn2≥ka2
so we have k≤∥x∥2/a2
number of xi with ∥xi∥≥a is no more than ∥x∥2/a2
In terms of RMS value:
fraction of entries with ∣xi∣≥a is no more than (arms(x)2)
e.g., no more than 4% of entries can satisfy ∣xi∣≥5rms(x)
Distance
Euclidean distance between n-vectors a and b is
dist(a,b)=∥a−b∥
agrees with ordinary distance for n=1,2,3 rms(a−b) is the RMS deviation between a and b
Triangle inequality
Triangle with vertices at positions a,b,c
edge lengths are ∥a−b∥,∥b−c∥,∥a−c∥
by triangle inequality
∥a−c∥=∥(a−b)+(b−c)∥≤∥a−b∥+∥b−c∥
Standard deviation
for n-vector x, avg(x)=1Tx/n
de-meaned vector is x~=x−avg(x)1(soavg(x~)=0)
standard deviation of x is
std(x)=rms(x~)=n∥x−(1Tx/n)1∥
std(x) gives typical amount xi vary from avg(x) std(x)=0 only if x=α1 for some α
greek letters μ,σ commonly used for mean, standard deviation
a basic formula:
rms(x)2=avg(x)2+std(x)2
The Core Identity
The statistical identity relating the Root Mean Square (RMS), average (avg), and standard deviation (std) of a data vector x is given by:
rms(x)2=avg(x)2+std(x)2
This identity is derived from the geometric decomposition of a vector in an n-dimensional space, based on the Pythagorean theorem.
Vector Definitions
Let x be a data vector in Rn and 1 be the vector of ones in Rn.
x=x1x2⋮xn,1=11⋮1
We define the average (mean) of x as μ=avg(x)=n1∑i=1nxi=n11Tx.
The vector x can be decomposed into two fundamental components:
Average Component Vector:
A vector where each element is the mean, μ. This represents the constant part of the data.
μ1=μμ⋮μ
Deviation (De-meaned) Vector:
The vector of deviations from the mean. This represents the fluctuating part of the data.
x~=x−μ1=x1−μx2−μ⋮xn−μ
Orthogonal Decomposition
The decomposition of x is written as x=μ1+x~. The key geometric insight is that these two component vectors are orthogonal, meaning their dot product is zero.
Proof of Orthogonality:
(μ1)Tx~=(μ1)T(x−μ1)=μ1Tx−μ21T1
By definition, 1Tx=nμ and the dot product 1T1=n. Substituting these in:
=μ(nμ)−μ2(n)=nμ2−nμ2=0
Since their dot product is zero, the vectors are orthogonal: μ1⊥x~.
The Pythagorean Theorem & Final Derivation
Because the components are orthogonal, they form a right-angled triangle in Rn. The Pythagorean theorem applies to their squared norms (lengths):
∥x∥2=∥μ1∥2+∥x~∥2
The statistical terms are the mean of these squared norms. By dividing the entire equation by n, we derive the final identity. We use the definitions:
rms(x)2=n1∥x∥2
avg(x)2=μ2=n1∥μ1∥2
std(x)2=n1∥x~∥2
The final derivation is:
n∥x∥2rms(x)2=n∥μ1∥2+n∥x~∥2=avg(x)2+std(x)2
Mean return and risk
avg(x) is the mean return over the period, usually just called return.
std(x) measures how variable the return is over the period, and is called the risk.
Cheyshev inequality for standard deviation
For any two n-vectors a and b, the absolute value of their dot product is less than or equal to the product of their norms.
∣aTb∣≤∥a∥∥b∥
This is true because the geometric definition of the dot product is ∥a∥∥b∥cos(θ), and the absolute value ∣cos(θ)∣ cannot exceed 1. Written out in terms of their components, the inequality is:
The norm of the sum of two vectors is less than or equal to the sum of their individual norms. Geometrically, this means the length of any side of a triangle is less than or equal to the sum of the lengths of the other two sides.
∥a+b∥≤∥a∥+∥b∥
Proof
This inequality can be proven using the Cauchy-Schwarz inequality as follows.
∥a+b∥2=(a+b)T(a+b)=∥a∥2+2aTb+∥b∥2≤∥a∥2+2∣aTb∣+∥b∥2(since x≤∣x∣)≤∥a∥2+2∥a∥∥b∥+∥b∥2(by the Cauchy-Schwarz inequality)=(∥a∥+∥b∥)2
Taking the square root of both sides completes the proof of the triangle inequality.
∥a+b∥≤∥a∥+∥b∥
Derivation of Cauchy-Schwarz inequality
It's cleary true if either a or b is 0
so assume α=∥a∥ and β=∥b∥ are nonzero.
We have