## Vector and matrix multiplication

Don't jump to conclusions, and you will be fine. See Matrix notation and summation for the basic notation.

### Vector multiplication

**Intuition**. When I ask students how they would multiply two vectors, by analogy with vector summation they suggest

(1)

This is a viable definition and here is the situation when it can be used. Let denote the initial deposit (the principal) of client of a bank. Assuming that clients deposited their money at different times and are paid different interest rates, the bank now owes client the amount where the factor depends on the interest rate and the length of time the money has been held in the deposit. Thus, (1) describes the amounts owed to customers on their deposits.

However, one might ask how much in total the bank owes on all deposits, and the answer will be given by

(2)

**Definition**. (2) is called a **dot product** (because the dot is used as the multiplication sign) or a **scalar product** (because the result is a scalar). Although we provided a real life situation when (1) is useful, in fact there are deep mathematical reasons for using (2) instead of (1).

**Exercise 1**. Do you think that for any two vectors (**symmetry** of the scalar product)?

### Matrix multiplication

We start with a special case that gave rise to the whole matrix algebra.

**Global idea 2**. We want to write a linear system of equations

(3)

in a form .

Put , , Define the product by

(4)

Then, taking into account the definition of equality of two matrices, we see that (3) is equivalent to

(5)

**Digesting definition (4)**. Denote the rows of Then can be written as Such a representation of a large matrix in terms of smaller submatrices is called **partitioning**. Then (4) shows that elements of the product are dot products of vectors

(6)

Note that this definition is correct because the number of elements of is equal to the number of elements of Alternatively, the number of columns of equals the number of rows of

### General definition

It consists of two parts:

**Part 1** (**compatibility rule**, or **rule for counting dimensions**) (the number of columns of equals the number of rows of ).

**Part 2** (**rows by columns rule**, or **rule for multiplication**) Let us partition into rows and into columns:

Then the elements of the product are found as dot products of the rows of by columns of :

In words: to find the elements of the first row of , fix the first row in and move right along columns of .

Cayley appears to be the inventor of this rule, although many other parts of matrix algebra had been discovered before him.