A matrix right-multiplied by a column vector---that is, $\mathbf{A}\mathbf{x}$---is a linear combination of the columns of $\mathbf{A}$, with coefficients of the combination given by the entries of $\mathbf{x}$. For example,
A = round(5*rand(3,3))
x = [-1.;1.;2.]
3-element Array{Float64,1}: -1.0 1.0 2.0
A*x
3-element Array{Float64,1}: 2.0 11.0 12.0
x[1]*A[:,1] + x[2]*A[:,2] + x[3]*A[:,3]
3-element Array{Float64,1}: 2.0 11.0 12.0
In this context it's clear why the number of entries in $\mathbf{x}$ has to be the same as the number of columns in $\mathbf{A}$.
If you left-multiply a matrix by a vector, it should be a row vector. I'll be using column vectors exclusively, so we have to transpose (or conjugate transpose) to get the row shape. Transposition of the product also makes the interpretation clear:
$$ \mathbf{y}^* \mathbf{A} = \bigl( \mathbf{A}^* \mathbf{y}\bigr)^*, $$which is the (conjugate) transpose of the columns of $\mathbf{A}^*$, i.e., the rows of $\mathbf{A}$.
Linear algebra, including abstract linear algebra, is built in large part on linear combinations. The vectors that one combines need not be ordinary vectors in $\mathbb{C}^n$. For instance, a polynomial is a linear combination of monomials. We can encompasss this fact too by generalizing the notion of a "matrix":
$$ c_0 + c_1 t + \cdots c_n t^n = \begin{bmatrix} 1 & t & \cdots & t^n \end{bmatrix} \begin{bmatrix}c_0 \\ c_1 \\ \vdots \\ c_n \end{bmatrix}. $$The "matrix" here is sometimes called a quasimatrix. From here it's a small step to imagine choosing many values of $t$ in an interval and converting the quasimatrix into a proper matrix. This is easy to carry out in MATLAB.
t = linspace(0,1,51);
V = zeros(length(t),4);
for j = 0:3
V[:,j+1] = t.^j;
end
using PyPlot
plot(t,V)
4-element Array{Any,1}: PyObject <matplotlib.lines.Line2D object at 0x31d722150> PyObject <matplotlib.lines.Line2D object at 0x31d722350> PyObject <matplotlib.lines.Line2D object at 0x31d722490> PyObject <matplotlib.lines.Line2D object at 0x31d7225d0>
It's often useful to think of a matrix--matrix product as a collection of matrix--vector products. For example,
A = round(8*rand(4,4)); B = triu(ones(4,3));
C = A*B
4x3 Array{Float64,2}: 4.0 7.0 9.0 3.0 4.0 5.0 5.0 9.0 13.0 8.0 11.0 18.0
[ A*B[:,1] A*B[:,2] A*B[:,3] ]
4x3 Array{Float64,2}: 4.0 7.0 9.0 3.0 4.0 5.0 5.0 9.0 13.0 8.0 11.0 18.0
Hence the number of rows in $\mathbf{B}$ has to be the same as the number of columns in $\mathbf{A}$. In the special case where that number is 1, then $\mathbf{A}$ is a column vector and $\mathbf{B}$ is a row vector, and the result is called a vector outer product.
u = [ 1; 2; 3 ]; v = [ 1im -1im 1 ];
size(u), size(v)
((3,),(1,3))
u*v
3x3 Array{Complex{Int64},2}: 0+1im 0-1im 1+0im 0+2im 0-2im 2+0im 0+3im 0-3im 3+0im
The rank and inverse of a matrix are of critical importance in linear algebra. It's important to be aware, however, that they are not robust to perturbations. For example, it should be clear that this matrix is singular and of rank 1.
A = [ 0 1; 0 0 ]
rank(A)
1
However, any perturbation of the second row, no matter how small, mathematically changes the rank to 2 and thus makes the matrix invertible.
B = A + [0 0; 1e-12 0 ]
rank(B)
2
In numerical computation, representation of and arithmetic with real numbers cannot be exact. Therefore, the notions of rank and invertibility will have to be carefully reconsidered. In fact, it's questionable as to whether you should ever compute a matrix rank in finite precision.
In linear algebra one learns that a matrix is the expression of a linear transformation relative to a particular basis. Unless stated otherwise, when we write out the numerical values of a matrix, we have chosen the standard basis, whose elements are the columns of an identity matrix.
Say we have a basis whose elements, when expressed in the standard basis, are the columns of a square matrix $\mathbf{A}$. Let a vector $\mathbf{v}$ be given as coordinates $x_1,x_2,\ldots$ in that basis. Then we can convert the coordinates of $\mathbf{v}$ to standard by matrix--vector multiplication:
A = round(10*rand(3,3)); x = [ -1; 1; 2 ];
v_standard = A*x
3-element Array{Float64,1}: 18.0 9.0 17.0
Consequently, we can convert a vector from standard coordinates to the "$\mathbf{A}$-basis" through multiplication by $\mathbf{A}^{-1}$.
v_Abasis = A^(-1) * v_standard
3-element Array{Float64,1}: -1.0 1.0 2.0
Note that while A^(-1)*
works, it is not recommended, for reasons we will go into later. A mathematically equivalent but computationally preferable syntax is to use A\
instead.
v_Abasis = A \ v_standard
3-element Array{Float64,1}: -1.0 1.0 2.0