06-reference

3blue1brown linear transformations matrices chapter 3

Sun Apr 19 2026 20:00:00 GMT-0400 (Eastern Daylight Time) ·reference ·source: 3Blue1Brown (YouTube) ·by Grant Sanderson
3blue1browngrant-sandersonlinear-algebralinear-transformationsmatricesmatrix-vector-multiplicationbasis-vectorsi-hatj-hatmathematical-pedagogycanonical-explainerml-prerequisitesgeometric-intuitionessence-of-linear-algebratransformer-prereqneural-network-prereq

3Blue1Brown — Linear transformations and matrices | Chapter 3, Essence of linear algebra

Why this is in the vault

Chapter 3 of Essence of Linear Algebra (6.7M views as of April 2026, posted August 2016) is — in Sanderson’s own opening claim, and one I’d bet on — “the one topic that makes all of the others in linear algebra start to click.” The video keeps because (1) it provides the canonical anti-memorization framing of matrix-vector multiplication — every matrix is a linear transformation of space; the columns are where the basis vectors land; the multiplication formula (ax+by, cx+dy) falls out of “the transformed vector is the same linear combination of the transformed basis vectors as the original was of the original basis.” This single reframe converts matrix algebra from a memorized recipe into a geometric inevitability, and it is the most-cited pedagogical move in the whole Essence series; (2) it gives the operational definition of “linear” that actually transfers — grid lines remain parallel and evenly spaced, origin remains fixed — which is the cleanest test for distinguishing the special class of transformations linear algebra studies from the general class of functions that don’t decompose nicely; (3) it is the load-bearing prerequisite for every downstream concept in modern AI — neural network weight matrices, transformer attention’s Q/K/V projections, embedding-space rotations, dimensionality reduction, the entire vocabulary of “this matrix does X to your input” — none of which are coherent without Chapter 3’s geometric reframe; (4) the closing claim “every time you see a matrix you can interpret it as a certain transformation of space” is the highest-leverage single mental-model upgrade we can hand a Sanity Check reader who has only ever seen matrices as bookkeeping for systems of equations.

Core argument

  1. A transformation is a function from vectors to vectors, visualized as movement. The word “transformation” rather than “function” is deliberate — it primes you to imagine every input vector smoothly moving over to its output vector, which is the right cognitive frame for what’s coming.
  2. Visualize transformations as moving every point in an infinite grid. Keeping a faded copy of the original grid in the background lets you see where things ended up relative to where they started. Sanderson’s animation discipline is the visual-language moat here.
  3. Linear transformations are a special class with two visual properties. All lines remain lines (no curving), and the origin stays fixed. Equivalently: grid lines remain parallel and evenly spaced. Counter-examples (curving transformations, transformations that move the origin, transformations that look linear on horizontal/vertical lines but curve diagonals) sharpen the definition.
  4. A linear transformation is fully determined by where it sends i-hat and j-hat. Because grid lines remain parallel and evenly spaced, any vector v = xi-hat + yj-hat lands at x*(transformed i-hat) + y*(transformed j-hat). You don’t need to track every point — just the two basis vectors. Everything else is implied.
  5. Numerical formula falls out without memorization. If transformed i-hat = (a, c) and transformed j-hat = (b, d), then transformed (x, y) = x*(a,c) + y*(b,d) = (ax+by, cx+dy). This is the matrix-vector multiplication formula, derived geometrically rather than asserted.
  6. A 2x2 matrix is just packaging for two basis-vector landing spots. First column = where i-hat lands; second column = where j-hat lands. The matrix is shorthand for the geometric fact, not the fact itself.
  7. Worked examples make the framing concrete. 90° counterclockwise rotation: i-hat → (0, 1), j-hat → (-1, 0), so the matrix is [[0, -1], [1, 0]]. Shear: i-hat fixed at (1, 0), j-hat → (1, 1), matrix [[1, 1], [0, 1]]. Linearly dependent columns: 2D space squishes onto a 1D line — the geometric meaning of rank deficiency without the word “rank.”
  8. Universal closing thesis. “Every time you see a matrix, you can interpret it as a certain transformation of space.” Determinants, change of basis, eigenvalues, matrix-matrix multiplication — all become easier once this reframe is internalized. Chapter 3 is the inflection point of the whole series.

Mapping against Ray Data Co

Pedagogical structure (reusable template)

  1. Open with a thesis claim about the topic’s importance. “This is the topic that makes the rest click.” Sets the reader’s attention level.
  2. Parse the term. “Linear transformation” = “linear” + “transformation”; transformation = function but visualized as movement. Defining vocabulary up front.
  3. Establish the visualization regime. Move every point in an infinite grid; keep a faded copy of the original to anchor the eye.
  4. Restrict to the special class with two visual properties. Lines stay lines, origin stays fixed → grid lines parallel and evenly spaced. Use counter-examples to sharpen.
  5. Reduce the problem to the basis vectors. Because of the parallel-and-evenly-spaced property, knowing where i-hat and j-hat land tells you everything.
  6. Derive the numerical recipe geometrically. Don’t assert matrix-vector multiplication; show that it falls out of the basis-vector tracking.
  7. Package the basis-vector landings as a matrix. Reveal that the algebraic object is just shorthand for the geometric fact.
  8. Worked examples in both directions. Geometric description → matrix (rotation, shear). Matrix → geometric description (the [1,2],[3,1] puzzle). Show degenerate case (linearly dependent columns → squishing onto a line).
  9. Close with the universal interpretive claim. “Every matrix is a transformation of space.” Plant the flag for everything that follows.

Notable quotes

Source provenance