This is the first video in a trilogy aimed at demystifying the Laplace transform, a powerful tool for studying differential equations. Although we won’t dig into the Laplace transform itself until the next two chapters, everything that we cover here sets up the mental frameworks and the prerequisite knowledge that make understanding that transform as easy as I know how. This video is a lot more than just preamble, though. It’s a very fun lesson in its own right about how one of the most famous equations in all of math is enables a bizarre trick for solving an equation that is used ubiquitously throughout physics. The main characters throughout this chapter and the next two are exponential functions, and I’m always going to be writing these as e^(st). Here we think of t as being time and then s as just some number determining which specific exponential we’re talking about. One of the big aims of this video is to motivate using physics why it’s useful to give s the freedom to take on not only real number values, but complex ones as well.
[00:01:02] But wait a minute, what does it even mean to shove a complex number into an exponent? Here I imagine there’s a bit of a divide in the audience. There’s some regular viewers of math videos online for whom the specific case of plugging in π times i is a little bit cliched by this point. It is amply covered by many videos on YouTube, but on the other hand, most students find this to be an understandably baffling notion. Given how absolutely fundamental this is to everything that follows, even if you fall into that first camp, I hope you’ll agree that it’s worth kicking things off here by reviewing a very beautiful and visual way to understand what this idea is all about. The nice part is that what follows doubles as a gentle warm up for visualizing and thinking about differential equations. You start with the fact that e^t is its own derivative. And really, you should think of this as what defines the number e. Exponentials with other bases will have derivatives that are proportional to themselves,
[00:02:03] but e is the special number such that that proportionality constant is 1. Now, very often in a calculus class, you visualize derivatives as slopes of graphs. That’s all well and good, but it’s not the only way. You should get in the habit of flexing your mind a bit more. For example, let’s say you think of this as telling you the position of some point on the number line as a function of time. Then what the derivative expression is telling you is that at every moment the velocity vector must look identical to the position vector. And more specifically, because you know that e^0 equals 1, you also have an initial condition. It’s telling you you start at the number 1. So at the very first moment, the velocity is also 1, meaning it’s pointed to the right. But the farther to the right the position gets, the faster it must move. So even if you had never heard of the function e^t or exponential growth, this property alone is enough to give you a very visceral feeling for how it gives a value that grows and at an accelerating rate.
[00:03:00] But what if there was some constant in that exponent like e^(2t)? By the chain rule, the derivative is then two times the function itself. Reading this dynamically, it’s telling you the velocity vector is always two times the position vector. The growth gets out of hand all the more quickly. What if that constant was negative, say -0.5? The derivative is -0.5 times the function itself. So at every moment in time, that velocity vector looks like a 180 degree rotation of the position vector, but scaled to be half its length. This means you start moving to the left, but as you approach zero with a smaller position vector, that velocity gets proportionally smaller. So it approaches zero, but at an ever slowing pace. This, of course, is exponential decay. Now for the fun part: what if that constant was an imaginary number i, the square root of negative one?
[00:04:04] The chain rule tells us that the derivative is i times the function itself. Geometrically multiplying by i acts like a 90 degree rotation, so this is telling you velocity always has to be perpendicular to position. Recap on complex numbers: a complex value a + bi sits in the 2D plane. Multiplying by i: ai goes vertical (imaginary), and bii = b*(i^2) = -b. Each component is rotated 90 degrees, so the sum as a whole is rotated 90 degrees. Reading the equation: position is in the complex plane, and velocity is always a 90-degree rotated copy of the position vector. The only motion that satisfies this is rotation around a circle. The initial position is 1, so the velocity vector always has unit length — your point traces one unit of arc length per unit of time.
[00:05:02] For example: wait π units of time, you end up halfway around the circle — this is why e^(πi) = -1. Worth emphasizing how misleading the notation is. When you input a complex value, the expression really has very little to do with repeated multiplication, and honestly not that much to do with the number e. The actual computation is the Taylor series for e^x — plug in πi for each polynomial term: each extra factor of i rotates you 90 degrees, and the π^n / n! terms grow then shrink, and you get a spiraling sum that converges to -1. But focusing on the property of being-its-own-derivative is more helpful than focusing on the underlying computation. People get used to e^x as notational shorthand and think nothing of it.
[00:07:03] Throughout this lesson there’s a complex plane representing possible values for s. For each point on this plane, think about the corresponding exponential function e^(st). When s = i, the output traces a unit circle; the real part vs time is a cosine wave. If s = 2i, you rotate twice as fast. The convention: omega is the imaginary part and describes the angular frequency — radians per unit time.
[00:08:02] What if s has both a real and imaginary part — say -0.5 + i? You can split the exponential: the real part says the magnitude decays, the imaginary part says it rotates. Or alternatively: ask what combination of rotation and stretching would place the vector at 1 onto this new value s — for -0.5 + i, that’s a little over 90 degrees plus a small stretch. The derivative expression says the velocity always has that relationship to position — geometrically you can see the motion must spiral inward.
[00:09:00] Engineers call this the S-plane: each point encodes the entire function e^(st). Imaginary part of s tells you how rapidly it oscillates and which direction. Real part tells you whether magnitude grows or shrinks — positive real = exponential growth, negative real = exponential decay. The key equation telling us velocity is some modified version of position is basically a differential equation. Having seen how reading off a differential equation can explain what complex exponents mean (at least if you want e^x to retain its core property), let’s flip this around: a different set of differential equations can motivate why you would ever care about complex exponents.
[00:10:01] The central physics example: a mass on a spring. Position x changes over time. x = 0 is equilibrium. The first derivative is velocity, the second is acceleration. The key spring fact: the more you pull the spring to one side, the more strongly it accelerates the mass back. Force = mass × acceleration (Newton’s second law) is well-approximated as -k × position. k is a positive proportionality constant — the spring’s stiffness. Force is proportional to (negative) position.
[00:11:00] Often there’s also a term proportional to velocity — the damping term. Greek letter mu is another positive coefficient. The negative sign means: faster motion = stronger damping force. Could be friction, could be air resistance. Neither friction nor air resistance literally behaves like this, but it’s a first-order approximation of whatever the slowing forces might be. We now have a differential equation. The position over time is unknown, but it has to adhere to this constraint. Physical intuition: oscillation back and forth that decays as the system loses energy. Move everything to one side, set equal to zero. Reminder: with differential equations, there’s not one specific solution per se — different initial positions give distinct solutions. Initial position and initial velocity are the two free parameters. Every (initial-position, initial-velocity) pair corresponds to a distinct solution function. So to “solve” the equation is to find a family of functions, and ideally to know how to pick the family member that matches your initial conditions.
[00:13:02] The bizarre trick: simply guess that the answer looks like e^(st), where s is some constant to be solved for. As a calculus student this really bothered me — guessing and checking like this feels like asking the student to know the answer ahead of time. Also, physical intuition says exponential is not how a mass on a spring behaves. And yes, this is unsystematic. But the desire to make this trick more systematic and generalizable is one of the things that leads to the Laplace transform. If x(t) really did look like e^(st), then x’ = s · e^(st), x” = s^2 · e^(st), and all the other constants come along for the ride. Factor out e^(st), and everything that depends on time is in that one term.
[00:14:00] Exponentials never equal zero, so the rest of the expression must equal zero. What’s left is a piece of algebra — solve a quadratic in s, which looks like a mirror image of the original differential equation. Easiest case: ignore damping (mu = 0). With rearrangement and a square root, s = ± sqrt(-k/m). Both k and m are positive, so i (the square root of -1) has now entered the game whether you wanted it or not. Call sqrt(k/m) = omega. We were exploring the possibility that x(t) = e^(st); plugging in the values of s, you now know what that means.
[00:15:00] Plugging in a purely imaginary term corresponds to oscillation in the complex plane. On the one hand: weird, because the mass on a spring needs a real-valued solution, not these complex functions. On the other hand: oscillating matches what you want, and matches it quantitatively. Increase k (stronger spring), omega goes up, faster oscillation — physical intuition agrees. Even still: bizarre. The position of the mass on a spring is clearly a real number. Zoom out: for the pure mathematical equation divorced from physics, there exists a complex-valued function solving it, namely e^(iomegat). To squeeze out a real-valued solution you can either ignore the imaginary part (the animation hint) or — better, and the path that leads to Laplace — add up the two distinct complex solutions you found. When you add the two rotating vectors tip to tail, the result stays on the real number line. The way it oscillates over time looks like 2cos(omegat).
[00:17:00] Why are you allowed to add two solutions and get another solution? Because the equation is linear: if two distinct functions solve it, their sum solves it. More flexibility than that: if you scale each function by some constant and add, the scaled sum is also a solution. Solving the equation = finding the family of all solutions. The math came back with two distinct functions; because it’s linear you can scale each one and add them to get a valid solution. Scaling coefficients don’t have to be real — they can be complex, which affects the initial angle of each rotating vector. The family of all such scaled-and-added solutions IS the family of all solutions to the equation. Most are complex-valued; the real-valued ones are a special case, and which one you want depends on initial conditions.
[00:18:01] For example: initial position 2, initial velocity 0 — set both coefficients to 1, you get the sum we found earlier. Different initial position — scale both constants by the same amount. Objection: if this is supposed to motivate complex exponents, you could complain — needlessly complicated! If the strategy is to just guess some function with a free parameter, why not guess cosine or sine directly? Indeed: cosine OR sine totally solves this equation as long as the frequency is sqrt(k/m). Because it’s linear, you can scale both cosine and sine and add them — the family of solutions in an alternate coordinate system. Arguably way more sensible. All real solutions correspond to real scaling coefficients. Why complicate things with complex numbers?
[00:19:03] The value of putting exponentials front and center makes itself clear as soon as we try to generalize. When we solved for s without damping, we got two values constrained to the imaginary axis. Different k and m → different imaginary values → different oscillation frequencies. But what if we reintroduce damping mu, setting it nonzero? Solving the equation looks like applying the quadratic formula. The salient feature: the two solutions s now have not only an imaginary but also a negative real component. Just minutes ago we discussed exactly what e^s with negative real and imaginary part looks like — both decays and oscillates.
[00:20:01] Graph the real component of that exponential — it matches what you’d expect from a damped spring. Increase mu enough and eventually the solutions for s have no imaginary part — just decay. The spring is overdamped. This whole example is the damped harmonic oscillator — fundamental throughout physics. How far does this trick generalize? For any equation of the form: a bunch of higher order derivatives, each scaled by some constant, summed, set equal to zero — substitute e^(st), each derivative term picks up a factor of s, factor out the exponential, and you’re left with a polynomial in s that must equal zero.
[00:22:02] The fundamental theorem of algebra: polynomials always factor into linear terms exposing n roots, as long as roots can be complex. So for a fifth-degree equation you might get five roots in the s-plane. The math is telling you: e^(st) is a valid solution as long as s is one of these roots. Linear, so scale each exponential and add them — the family of all solutions. Scaling constants are knobs you can tune; they can be real or complex, influencing both amplitude and phase. Specific values depend on initial conditions. Glossing a nuance about repeated roots, but this is the general idea. Most real-world equations are not simple linear ones. The damped harmonic oscillator with an external driving cosine force is the classic next case — which came up in the optics video about why light slows down in a medium like glass (giving the prism effect).
[00:23:03] In that scenario, charges inside glass were modeled as little damped harmonic oscillators being driven by an incoming light wave (a sine of some frequency). The frequency of incoming light has nothing to do with the natural resonant frequency of the oscillator. The same equation as before, with an added cosine driving term. The family of solutions is no longer a simple linear combination of exponentials with freely-tunable coefficients. The dumb e^(st) trick does NOT work directly. But everything we’ve discussed brings you a lot closer than you might expect. The driven solution does happen to look like a combination of four specific exponentials — it’s just that you can’t freely tune all the coefficients, they’re constrained.
[00:24:01] The whole substance of the prism example comes down to understanding exactly how big these coefficients are as a function of the incoming light frequency. A surprisingly common outcome: solutions to differential equations from the real world look like a combination of exponentials but with particular coefficients. This ubiquity of exponentials is why engineers benefit from intuitive understanding of points on the S-plane and how they encode growth, decay, and oscillation. You can think about e^(st) as the atoms of calculus. Complicated functions describing our world can often be broken into these parts, as long as you give s the freedom to be complex. The more you allow infinite combinations — potentially over a continuum of values for s rather than a discrete set — the more powerful this becomes.
[00:25:02] The key question: given some unknown function and a differential equation describing it, even if you assume it can be broken into exponential parts, how do you actually find what those parts are? For the forced harmonic oscillator: how would you know the solution is built out of four specific exponentials? How would you solve for the s values and the corresponding coefficients for a particular initial condition? There is a tool for this job, and it’s something known as a Laplace Transform. If you watched the Fourier series chapter, this is ringing bells: imaginary exponentials describing rotation, breaking up general functions as a sum of those exponentials. Absolutely a connection — the Laplace transform extends Fourier series and Fourier transforms to a much more general family of functions.
[00:26:04] High-level preview: when you use a Laplace transform to solve a differential equation, it ends up looking remarkably similar to the dumb trick of substituting e^(st). With the dumb trick, the differential equation turned into algebra basically because taking a derivative is the same as multiplication by s for these specific functions. The Laplace transform does the same thing for the same reason. It translates functions into a new language where e^(st) — the atoms of calculus — are the fundamental units.
[00:27:01] In this language, derivatives look like multiplication, and differential equations look like algebra. Next chapter: how the transform is defined, how to visualize what it’s doing, how to use it to concretely solve a nonlinear (driven) equation.