Link back to Home page
Link to typed notes:
0. Complex Numbers
Why is i on the other side??
Graphically,
Definition Modulus of a complex number
Think of it like vectors and distances from origin
Definition Argument of a complex number is the angle between
The argument lies in the interval
Polar form
Every non-zero complex number can be written in the form of
Where
Let
Multiplication
Division
Powers (De Moivre's)
Roots of units
Example
Take
- Modulus =
- Argument =
$ - Polar form:
- Exponential form:
Conjugates and Reciprocals
Why do they matter?
- Getting rid of the
in the denominator - Long-standing convention that we want denominator to be real numbers
- Modulus and argument can be easily seen
- Calculating the magnitude
- Later - calculating products in vector spaces
- Roots of real polynomials as they come in conjugate pairs
If
-
This is essentially reflecting
is the real axis -
If
then $$ z \bar{z} = \mid z \mid
The reciprocal by definition is the number that when multiplied to gives the result of
In terms of geometry,
Reciprocals: $z^{-1} = \frac{\bar{z}}{\mid z \mid
{ #2}
}$
Proof:
Let
We want
Therefore:
The conjugate
And the modulus is
Therefore $$ z^{-1}= \frac{\bar{z}}{\mid z \mid
{ #2}
} , \space z \neq 0$$
Reciprocals in polar form
Any non-zero complex number can be written as:
The reciprocal
But
So we have
This form can also be derived from the earlier formula:
We have
Polar form is very useful when converting numbers from their cartesian form to mod-arg form
If we had
But if we had reciprocals in polar form with
Consider this example:
- Cartesian way:
And now consider the polar way:
Proof of fundamental theorem of algebra here
- Need intermediate value theorem to prove stuff
Roots and conjugate pairs
If a polynomial has real coefficients, then any non-real complex root must appear with its complex conjugate as another root
So if
Why?
Suppose the polynomial is
Now assume that
This is essentially saying that if we plug is
Then if we take the conjugate of both sides:
However,
And because the coefficients are real:
So we've shown that $$ P(\bar{z}) =0 $$
Hence, if
Solving roots in polar form
1. Real space
Definition The real n-space
Eg:
Etc…
Scalar/Dot product
A scalar product (aka dot product) of two vectors returns a scalar (i.e a real or a complex number)
For
Properties of scalar product
- Commutative
- Distributive
Definition Orthogonal vectors - vectors at right angles
Norm of a vector
Norm is basically the magnitude or length of a vector
Unit vector - A vector with the magnitude of 1
Generalized Pythagoras theorem:
Projections
Intuitively, imagine we have two vectors:
The projection of a vector
Now, the problem is,
We know that any point on the line
We need to find a
and the vector from
Since perpendicular vectors have dot product = 0, we have
Therefore, the component of
Cauchy-Schwarz Inequality
Using the dot product definition, we have
Since
This is the Cauchy-Schwarz inequality, and it essentially means that a project (shadow) can never be longer than the actual vector.
Triangle inequality
The triangle inequality here below can be proved using the Cauchy-Schwarz inequality:
After expanding the LHS, we can how apply the Cauchy-Schwarz inequality
So we have
After taking square roots, we have:
Equation of a Line in
To define a line, we need two things (maybe more, but two main):
-
A point the line passes through, called the position vector or the point of origin of the line. Call it
. So if the line passes through a point then -
A direction the line travels in, called the direction vector. Call it
There are a few ways to describe lines.
Vector equation
The vector equation of a line is the most geometric form, if the line passes through the point
where
Parametric equation
This is just a component form of the vector equation, but it's useful when working with intersections or substituting into other equations for
If we write vectors in components:
Then
This gives
In this case, the parametric equation of a line is:
Cartesian equation
The cartesian form eliminates
Converting from parametric to Cartesian equation example
When converting from parametric to cartesian, our goal is to eliminate
So given our three parametric equations above, we can solve each of them for
Geometrically speaking, we get a ratio of how much we've moved along the line (scaled by
Parallel, Intersection, Skew
In
It's useful to outline what each of them mean and what their test would look like:
- Coincident - the direction vectors are multiple of each other as one point satisfies the other line's equation
- The points lie on the same line
- Parallel - direction vectors are multiples, but there is no common point
- The lines are in the same direction, but different positions
- Intersecting - system has a single consistent solution for the parameters (
) - The lines meet a single point
- Skew - system has no solutions (inconsistent equations)
- Not parallel, never meet
- Lie in different planes
Forming and solving systems of equations to determine the arrangement of lines in
Let the two lines
where
are position vectors of points on each line. are the direction vectors are scalar parameters
In component form, this gives a system of 3 linear simultaneous equations in the two unknowns
And we we can either
- Eliminate one parameter
- Use the determinant of matrix (more on this later)
- Substitute using pairs of equations
Once we have our result, we can use our definitions to check the arrangement of lines in
- We can only have a unique point of intersection if the system has a unique consistent solution for
- If the system has no solution and are not parallel, they are skew - i.e. lie in different planes
Equation of a plane
Let's start with the intuition, what is a plane? (Yes it is something that flies..)
A plane is a flat 2D surface that extends infinitely in three dimensions. We can think of it as :
- The set of all points that satisfy one linear condition in
- Or, all points that we can reach starting from one point and moving along two independent direction vectors
Definition A plane can be defined by a point
Vector equation
So a plane is a set of all points
so if
Check that this makes sense, the dot product being zero means that the vectors are perpendicular
Cartesian equation
Say the normal vector
Where
Also note that $$ \vec{r} \cdot \vec{n} = \vec{a} \cdot \vec{n} $$
Intuitively, this equation defines all the points
There are pretty good 3b1b videos on Linear Algebra for some visual understanding. I often imagine solving problems in Grant Sanderson's voice to give myself some confidence when approaching questions
Parametric equation
As with lines, planes can also be defined using a point and two non-parallel direction vectors
So if
And every point on the plane is reachable by some combination of
Vector product
To find perpendicular/orthogonal vectors, we need to have something that lets us calculate that, and this is where the vector product comes in. The ’trick’ so to speak of calculating this is similar to finding determinants in matrices
Let
be two vectors in
The vector product
Vector product properties
- Vector products are NOT communicative and NOT associative
When we take the cross product
Parallelograms...
Where
Given that there are two directions to a plane, up and down, we need to know which direction the cross product points towards. And this is where the right hand rule comes in.
- If we point the index finger along vector
- The middle finger along vector
which is at right angles to - Then our thumb now points in the direction of
Now, if we flip the order, and instead do
So we get:
Geometrically speaking, the cross product represents the oriented area of the parallelogram spanned by
Intersection of Lines and Planes in
There are 3 cases for an intersection between a line and a plane:
- No intersection -
is parallel to - Infinitely many points -
is contained in - One point
intersects in a single point
The first two cases are pretty straightforward. And the final case can be checked by computing the normal vector to
Now, the intersection of 3 planes in
- The intersection is a single point
- The intersection is a line (star)
- The intersection is another plane when all planes coincide
- The intersection is empty (at least two planes are parallel or they form a prism)
Distances in
Point and a Plane
Given a point
Say if we've got a plane:
and a point
Then the distance from
Imagine standing above a plane with a flashlight pointing along the normal vector. Then the shadow of the position vector onto that normal is the perpendicular, i.e the shortest path.
Point and a line
Let's say we want to find the distance between the line
Use Pythag, we can show that the distance between
We can do this by considering projections. The projections of
The
The distance is therefore given by the norm or the magnitude of both sides
2. Matrix Algebra
Definitions
Definition Suppose that
An
The coefficients
For any
Definition Zero matrix
Size
Definition Diagonal matrix
A square matrix
(i.e., all non-diagonal entries are zero):
Definition Identity matrix
An
Thus
The addition and subtraction for matrices is only defined for matrices of the same size, but multiplication, as well will see below, has other properties
Why can’t we add or subtract matrices that aren’t the same size??
Matrix multiplication
Let
then
The matrix product is essentially the product of the row vectors in
So,
So the product
Properties:
- Distributivity
- Associativity
Transposition of a Matrix
Let
In other words, rows of
Example:
If
Properties
- Flipping twice returns the original matrix - Transposing a sum is the same as sum of the transposes - a scalar doesn't affect a transpose - when we flip a product, we reverse the roder
Inverse of a Matrix
For numbers, the inverse "undoes" the effect. Eg: $$ 3 \times \frac{1}{3} = 1 $$
For matrices, the inverse works in the same way, it undoes the transformation a the matrix does:
So $$ AA^{-1} = I$$
Where
- Only square matrices can inverted, but not all square matrices are invertible
- An inverse exists iff
An
where
is the
Let
Then
Properties
Let
has a unique inverse is invertible and is invertible and is invertible and
Powers of a Matrix
If
- If
and commute, ( ) then
3. Systems of Linear Equations
A system of
And in matrix form:
Where:
= coefficient matrix - the column vector of variables
Augmented Matrix
The augmented matrix of system is
Row Operations
There are 3 allowed row operations that allows us to simplify systems without changing the solution set. I.e. matrices are are row equivalent
- Row operation are invertible, i.e. their inverses are also row operations
Row Scaling
Row interchange (swap)
Row replacement
Gaussian Elimination
Consider we're trying to solve a system of simultaneous linear equations like
Instead of solving one variable at a time using substitution, we can turn this into an augmented matrix.
So we instead have:
So each row is an equation, and each column, except the last one, is one variable's coefficients.
The idea now is to use the row operations listed above to reduce the augmented matrix to row echleon form (REF) or reduced row echleon form (RREF)
A pivot is the first non-zero entry in a row of a matrix
A matrix is is REF if:
- Every pivot is to the right of all pivots above it
- Zero rows are at the bottom
A matrix is in RREF if:
- It’s already in REF
- Every pivot is
- Every column containing a pivot has all entries equal to
Definition Gaussian Elimination is the method of solving linear equations by starting with
Each row in the matrix represents a plane in 3D (or a line in 2D or a hyperplane in higher dimensions). So Gaussian elimination is essentially rotating and scaling these planes until they're "aligned with the axes". The intersection point doesn't change, we just manipulate it to make it easier to understand
Reading solutions
Once our matrix is in REF, the augmented matrix will correspond to the solutions of the equation
Reducing a matrix to RREF
Our goal is to transform a given matrix
- Each row starts with a
(pivot) which is to the right of the pivot above it - Each pivot column has zeros above and below the pivot
- Any rows of all zeroes go to the bottom
The means that:
- Each pivot correspond to a leading variable
To reduce a matrix to RREF:
- Starts with the first non-zero column going from left to right - this will be our pivot column
- If the first entry is non-zero, use that, otherwise, swap that row with one below so that the pivot moves to the top of the block
- Make the pivot
using row operations - Use row operations to make all entries below the pivot =
- Now move diagonally (one column to the right and down) and repeat the process until all pivots have zeroes below
- And go back and make all entries above the pivot zero
Matrix inverses using row operations
Definition A system of linear equation is consistent iff the augmented matrix, when put into this REF has no rows of the form
Each matrix has a unique matrix in RREF
If
Given any
Here, the key idea is that row operations that turn
Definition Elementary matrix - A matrix obtained from the identity matrix
Each elementary matrix is invertible, and it’s inverse is an elementary matrix of the same type. Ie. if we multiply
If
is invertible - The unique solution to
is - The RREF of
is is a product of elementary matrices
If
Algorithm for finding inverses using row operations
Suppose we have an
We first form an augmented matrix:
Then we apply row operations to this matrix to being the left half
Since we can get from
Why the RHS equals
As each row operation can be represented by an elementary matrix
But in the augmented matrix form, we've been applying those same exact operations to the identity matrix on the RHS, so the RHS ends up being:
And therefore:
Example
Suppose we want to find the inverse of the matrix:
First we write it in the augmented form:
We want to scale so that the first pivot is 1, so we have:
Next, we want to eliminate below the pivot:
Now, as the LHS is
Rank of a Matrix
Definition Given a
The rank essentially measures how many independent rows or columns a matrix has. So for an
- No row/column can be written as a combination of others
- Every pivot position is filled
- So Gaussian elimination can turn
into so that no zero rows appear - This means we can find
Therefore, if
This is also why if there is a zero row in an RREF form, the inverse of a matrix does not exist
Properties of a Rank
Let
row equivalent to rank( ) = rank( ) - rank(
) = rank( ), - rank(
) - Based on the definition of rank, we can never have more than
linearly dependent vectors in Same with the rows. So as row rank = column rank, the inequality follows.
- Based on the definition of rank, we can never have more than
- rank(
) rank( ) + rank( ) - provided is - Rank(
) min(rank( ), rank( ))
4. Determinants
Definition
A function
1. Linearity in each row
If we multiply a row by a scalar
And if we add the rows, determinants add. So the sum of the determinants is the determinant of the sums
2. Alternating
If
So if two rows are equal, then the determinant is 0
Swapping two rows flips the sign of the determinant
3. Normalisation
Using the axioms above, we can further prove the row operations and their effects on matrices
- Row interchange doesn't change the determinant
- Swapping rows (row replacement) causes the determinant to be me multiplication by
as it flips orientation - Multiplying a row by a scalar multiplies the determinant by that scalar.
- So if
- So if
So
where
Determinants of Upper Triangular Matrices
An upper triangular matrix is a matrix where all entries below the diagonal are zero:
So if
Intuitively, the matrix
- Scales the first coordinate by
- Scales second coord by
- Third by
- And it may shear the shape horizontally because of the entries above the diagonal.
Each diagonal entry
A more formal proof can be done by using row operations and transforming the above matrix into
Determinant of any square matrix
Once we have the facts above, we now have an algorithm to compute the determinant of any square matrix:
To compute
- Perform Gaussian elimination to reduce
to an upper triangular form i.e. the REF for a square matrix - Keep track of how each row operation affects
- Compute the determinant of the upper triangular matrix
Notation wise, determinant of
Proof for a matrix
For
Geometrically speaking, in
- The sign of
tells us if the orientation is preserved or flipped. If the parallelogram collapses to a line (area = 0), then then two vectors are linearly dependent, i.e. the matrix is not invertible. - That's why
- In the similar way, the rank tells us what the shape "collapses" to in higher dimensions.
Now, for the more formal proof, we can compute
We can do the row operations
From the property of determinants, the row replacement does not change the determinant, so
Therefore,
The determinant of a matrix with integer entries is always an integer
Existence and Uniqueness
This will be proved in Linear Algebra II by giving a different definition of determinant. So the lesson is to never trust anything.
Determinants and Invertibility
Proof
Starting with the
If
Now we can take the determinant of both sides
This means that
Now for
For a bit of intuition, recall that the determinant is geometrically the scaling factor of volume when
Now for the algebraic proof. We know that when we apply Gaussian elimination to turn
Each
Now, if
But
There is another proof by contrapositive, given in the lectures notes.
So if we can prove that
So if
And if a matrix has a zero row, the determinant will be
Determinant using Cofactors
Now that we know how to calculate the determinant of a
So if
For some definitions, consider an
Then:
- minor
is the determinant of the smaller matrix we get after deleting the -th row and -th column. - cofactor
adds a sign to the minor:
So to find the determinant, we pick a row/column, say row 1, then for each element
- Delete its row and column from the matrix to get a smaller matrix called the minor
- Multiply that smaller determinant by
- Multiply by a sign
that alternates across the row
So formally, if
Then
Where
And
So the formula for any
The sign pattern comes from one of the determinant axiom which tells us that if we swap two rows, the determinant changes the sign.
As we're expanding the determinant along a row, we're sort of "isolating" one element
This is quite efficient when we have a matrix with lots of zero entries
Determinant of a Transpose
For any square matrix
What this is means transposing a matrix (i.e. flipping rows and columns) of a matrix does not change its determinant
Intuitively, if we think determinant as the volume scaling factor, then transposing it just changes our point of view, so instead of looking at how
The other way to think about is using the fact that the determinant is the product of the diagonals, and when we transpose a square matrix, the diagonals remain the same, hence the determinant is the same.
Determinant of a Product
If
Geometrically again, scaling a volume by a factor of
More formally speaking, we can think of every invertible matrix
And as each elementary matrix corresponds to a single row operation, and because we know how determinant changes under row operations:
The proof in the notes is bit more detailed and uses the 3 cases of the elementary matrices, but the idea is the same.
Powers of a matrix
The proof is rather trivial, and can be proved by induction if wanted as an exercise to the reader
Inverting a Matrix using Cofactors
This section is skipped as it will not be examined
5. Linear Transformations
This is where the proper linear algebra starts. A linear transformation is just taking a function from one vector space to another, following a couple of specific rules, namely additivity and scaling.
Definition Let
Intuitively, a linear transform never bends, curves, or shifts space, it can only scale, rotate, shear, reflect, etc. So essentially, if the origin doesn't stay fixed, then the transformation is not linear.
The most important thing is probably this: every linear transformation is a matrix multiplication.
So if
So every linear transform, is a multiplication by some matrix
Now that we know that
Consider the standard basis vectors in
Where the
And we know that any vectors
So
Hence just showing the basis vectors, is enough to define the entire transformation. It's like taking the unit square and seeing how it changes under a matrix.
Linear Transformation in
Every linear transformation in
So to understand a transformation, we need to think about what it does to
Rotation about the origin
Let
Now think about what happens to
And when
Therefore the matrix of rotation is given by:
We can check that rotation preserves the area, so the determinant of this matrix is in
And therefore the matrix of linear transformation is given by:
Reflection in a line through origin
Note that a reflection keeps points on the line fixes, and flips points perpendicular to the line. There is a diagram in the notes which makes it easy to see why its
A reflection through the angle
The matrix is therefore given by:
The determinant of this matrix, when calculated, is
Stretch/shrink
A scaling just has the effect of multiplying axes. So the matrix here is self-explanatory:
Shear
It keeps one of the axis fixed, and moves all the other points in that direction parallel to the axis.
Composition of Linear Transformation
Composition of transformations is the same as what it's like in functions:
So we apply
Matrix multiplication is the same as composing linear transformations. So if
So the matrix of the composite transformation
The composition of a linear transformation is a linear transformation. It can be proved using the definition of composition and the transformation definitions above
Finding linear transforms
Inverses of Linear Transformations
Definition A linear transformation
Like any inverse, an inverse transformation just undoes the original transformation.
Proof of the inverse of linear transforms
Think back to calculus, an inverse exists
So suppose
So let
And now we need to show that
So let
Consider
As
As
Now... for the scalar multiplication, consider
As
Using the injectivity of
Invertibility
Let
As this is an
So
So let
By composition of matrices,
So since,
And since
Therefore we have that:
And by definition
Now for the
We assume that
Let
So we know that
Now to compute the other composition to check for invertibility:
So
And since both compositions have given us the identity,
6. Subspaces of
Vector subspaces
Intuitively speaking, a vector subspace of
It has three properties:
- It must contain the origin (i.e. be non-empty)
- It's closed under vector addition
- Closed under vector multiplication
So anytime we're given a vector, we just need to check for the properties above to see if it's a subspace of
Null Spaces
Let
I.e. it's basically the set of all vectors that get mapped to the zero vector for some
The null space is a subspace of
- closed under addition
- Closed under multiplication
And by linearity of
Invertibility and Null spaces
If the only solution to
Linear Span
Linear span is just the set of all linear combinations of vectors, and this is always a subspace because linear combination rules automatically satisfy the subspace axioms
Range and Column Space
The column space is of a matrix
And the range is the subset of
And the range is also a subspace, which can be tested using the properties of subspace again
Linear Independence
Vectors are linear independent if:
What this means is that no vector can be written as a combination of others. So if a columns in the matrix are linearly dependent:
- The transformation keeps all directions and matrices are transformations
- There is no collapse of space
- So matrix is invertible
Trivial solution just means that the solution is
So with linear independence, if we can combine the vectors with non-zero coefficients to make them
So for a square matrix, all of the conditions below are equivalent:
- Columns independent
- column space =
being invertible
Test for linear independence
To see if vectors are linearly dependent or not, put the matrix in REF and look out for non-zero rows
Bases
A basis is just a minimal set of vectors that can generate (span) the whole space with no redundancy. So they give you all possible directions in your space, and the smallest possible set of directions. A bit like a choosing the axes of the space.
So in
Definition
A collection of vectors
- They span
- They are linearly independent
Basis are like the coordinate system for a vector space. The other important thing about basis is hat they are unique. Every vector
7. Eigenvectors, Eigenvalues and their applications
A bit of a diversion, as in my notes I will try and focus on the intuitive idea and then the definition.
So consider a linear transformation
Definition
A number
So eigenvectors are the directions a transformation preserves, and eigenvalues are the scaling factors for those directions.
Determinant criteria
Notice how the left hand side of the equation is vector multiplication, but the right side is scalar. So to turn the RHS into a vector multiplication, we can write
So we have:
And we can rearrange and factor out
When solving the equation above, all we're doing is looking for a non-zero vector that gets sent to zero by the linear transformation
The reason for this is because the kernal of
And this happens when the determinant is
So an eigenvalue exists when the sifted matrix
Characteristic polynomial
There’s a reason it’s called a characteristic polynomial, because it satisfies the characteristics of the matrix equation.
This basically helps us solve for values of
Upper triangular matrices
Say we have an upper triangular matrix:
And now consider the matrix
This is still an upper triangular matrix! And the determinant of this matrix is the product of its diagonal entries.
And if we set this equal to
Therefore, the diagonal entries of an upper triangular matrix are the eigenvalues of that matrix!
Applications in Google’s Page Ranking system
Read the MathsRant blog page!
Symmetric Matrices
A matrix
Let
has linear independent eigenvectors - All eigenvalues of
are real. - There exists an orthonormal basis of eigenvectors (more on this this later)
is diagonalisable by an orthogonal matrix
We can also think about this as the dot product with matrix product as scalar product
This is a result of Spectral Theorem, but the proof was omitted.
So as
We want to find linearly independent eigenvectors because a basis must be definition be linearly independent and geometrically, we need enough directions to describe the whole space. So if eigenvectors are dependent, they don't span enough directions.
Multiplicity
Multiplicity just tells us how many times an eigenvalue appears as a root of the characteristic polynomial. This is algebraic multiplicity.
Eg:
Then there is geometric multiplicity, which tells us the dimension of the eigenspace (number of independent eigenvectors).
The key rule is that
However for symmetric matrices, the geometric multiplicity
Orthogonality of eigenvectors - Thm 7.27
For symmetric matrices - eigenvectors corresponding to distinct eigenvalues are orthogonal.
Recall that two vectors being orthogonal means that they essentially meet at right angles and are in completely independent directions. So movement in one direction has no component in the other. So transformations by the eigenvectors stretches space along perpendicular directions for symmetric matrices.
Diagonalisation of a Matrix - Thm 7.30
At its core, diagonalisation is trying to answer a fairly simple question: can we find a coordinate system where the linear transformation acts in the simplest possible way? i.e. just stretching along axes without rotating or shearing?
So a diagonal matrix
Geometrically, each coordinate directions moves independently, and no direction interferes with another. Most matrices are not in this form because diagonalisation is about changing basis to one that the transformation works well with, and that basis being the eigenbasis
Recall that
So applying
Change of basis
Suppose that
Let
Then all
Now consider taking a vector in standard coordinates of a matrix
- Convert it into eigenbasis coordinates
- Apply the transformation in that basis to get a diagonal matrix
- Convert back to standard coordinates
Writing this as a composite transformation, we have that:
So finding a diagonal matrix just means finding perpendicular directions where the transformation acts by pure scaling.
Powers
Diagonalising also helps us computing large powers:
This is also why symmetric matrices also diagonalise, because they have enough linearly independent vectors
8. Orthogonal sets and Quadratic forms
Orthogonal vectors are essentially directions that don't overlap. So they're independent in the geometric sense. Think of the
Definition A set of vectors
Therefore, we can see that orthogonal vectors are automatically linearly independent, this will help us compute projections, eigenvectors, and simplify quadratic forms
Definition An orthonormal set is the same as orthogonal, but now each vector has a unit length, i.e. we have normalised the vectors
So a set is orthonormal if:
Orthonormal is more useful than just orthogonal because the coordinate along
So we just have to deal with dot products instead of worrying about scaling
Orthogonal matrices
Intuitively, an orthogonal matrix (thinking in terms of transformations):
- Rotates spaces
- Reflects space
- never stretches or squashes space
So the lengths and angles are preserved
Definition
A matrix
Then the following are all equivalent:
- Columns of
form an orthonormal basis
The reason for why
- Dot products of columns give the identity
- Diagonal has a unit length (1)
Gram-Schmidt Orthogonalisation
Before diving into the algorithm, it's good to think about the problem this process is trying to solve.
Consider vectors spanning a space, most of them are messy and angled, and what we want is the same space, but having all directions be perpendicular.
So the intuition behind Gram-Schmidt is given the vectors
- Keel
- From
, subtract the part pointing in the direction of (the projection) - From
, subtract the projection of - Normalise everything
So we remove the overlap of vectors step by step without changing the span.
I will not be covering the proof as that is given in the lecture notes, and my notes are just to cover a bit of the intuition that I find isn't in the notes, but it's also worth watching some YT videos to understand the intuition behind it. Tom Crawford has a nice video on it.
Orthonormal basis
A basis, as talked about briefly earlier, is just a coordinate system. Recall that for a set of vectors to form a basis of
- Span
- Be linearly independent
With that, an orthogonal basis is just a basis where all directions are perpendicular
Now for an important Theorem: Every subspace of
What this means is that no matter how tilted or messy as subspace is:
- We can always choose perpendicular axes inside it
- Geometry is always recoverable
This isn't obvious, but Gram-Schmidt process constructs it
Diagonalising Orthogonal Symmetric Matrices
Recall from previous chapter that diagonalisation is simply finding a coordinate system where the matrix acts simply.
For an orthogonal symmetric matrix,
- Eigenvalues are real
- Eigenvectors are orthogonal
Symmetric matrices also respect the dot product.
Quadratic Forms
A quadratic form is a function:
They are used to measure curvature and distance-like behaviour, so they describe shapes instead of transformations.
So every quadratic form becomes a sum of squares
Michael Penn has a good video on quadratic forms under the Number Theory playlist. Most of these notes, especially for the final two chapters were completed during the holidays, so the priority is given to doing past papers and building sufficient intuition instead of copying proofs into my own notes.
Definiteness
When working with quadratic forms, definiteness is a property of the sign of
We want to think of
There are 4 kinds of definite. So we let
Positive/Negative definite
This means that no matter which direction we go in, it is strictly positive/negative
Positive/Negative semi-definite
Never positive/negative, but can be zero is some directions
Indefinite
So the sign depends on the direction
Link to eigenvalues
This is also nice reason for why diagonalisation exists:
For a symmetric matrix, we have:
where:
are eigenvalues is written in the eigenvector basis
Now,
The lectures skipped the section on conic sections as that will not be assessed, and that concludes Linear Algebra I