NAVIGATION
Home
Research
Bookshelf
Garden
FIND ME ON
GitHub
LinkedIn
Email
Definition (Covariance matrix)
Given a random vector Xn=(X1,⋯ ,Xn)X^{n}=(X_{1},\cdots,X_{n})Xn=(X1,⋯,Xn), the covariance matrix of XnX^{n}Xn, KXi,XjK_{X_{i},X_{j}}KXi,Xj, is defined as follows KXi,Xj=(Var(X1)Cov(X1,X2)⋯Cov(X1,Xn)Cov(X2,X1)Var(X2)⋯Cov(X2,Xn)⋮⋮⋱⋮Cov(Xn,X1)Cov(Xn,X2)⋯Var(Xn))K_{X_{i},X_{j}}=\begin{pmatrix} \mathrm{Var}(X_{1}) & \mathrm{Cov}(X_{1},X_{2}) & \cdots & \mathrm{Cov}(X_{1},X_{n}) \\ \mathrm{Cov}(X_{2},X_{1}) & \mathrm{Var}(X_{2}) & \cdots & \mathrm{Cov}(X_{2},X_{n}) \\ \vdots & \vdots & \ddots & \vdots \\ \mathrm{Cov}(X_{n},X_{1}) & \mathrm{Cov}(X_{n},X_{2}) & \cdots & \mathrm{Var}(X_{n})\end{pmatrix}KXi,Xj=Var(X1)Cov(X2,X1)⋮Cov(Xn,X1)Cov(X1,X2)Var(X2)⋮Cov(Xn,X2)⋯⋯⋱⋯Cov(X1,Xn)Cov(X2,Xn)⋮Var(Xn)
Differential Entropy
Multivariate Gaussian
Principal Component Analysis