§ Discriminant and Resultant


I had always seen the definition of a discriminant of a polynomial p(x)p(x) as:
Disc(p(x))an(2nn)i<j(rirj)2 Disc(p(x)) \equiv a_n^{(2n - n)} \prod_{i< j} (r_i - r_j)^2

While it is clear why this tracks if a polynomial has repeated roots or not, I could never motivate to myself or remember this definition.
I learnt that in fact, this comes from a more general object, the resultant of two polynomials P(x),Q(x)P(x), Q(x), which provides a new polynomial Res(P(x),Q(x)Res(P(x), Q(x) which is zero iff P,QP, Q share a common root. Then, the discriminant is defined as the resultant of a polynomial and its derivative. This makes far more sense:


This cleared up a lot of the mystery for me.

§ How did I run into this? Elimination theory.


I was trying to learn how elimination theory works: Given a variety V={(x,y):Z(x,y)=0}V = \{ (x, y) : Z(x, y) = 0 \}, how does one find a rational parametrization (p(t),q(t))(p(t), q(t)) such that Z(p(t),q(t))=0Z(p(t), q(t)) = 0, and p(t),q(t)p(t), q(t) are rational functions? That is, how do we find a rational parametrization of the locus of a polynomial Z(x,y)Z(x, y)? The answer is: use resultants!



§ 5 minute intro to elimination theory.


Recall that when we have a linear system Ax=0Ax = 0, the system has a non-trivial solution iff A=0|A| = 0. Formally: x0    A=0x \neq 0 \iff |A| = 0. Also, the ratio of solutions is given by:
xi/xj=(1)i+jAi/Ajx_i / x_j = (-1)^{i+j} |A_i|/|A_j|

If we have two polynomials p(a;x)=a0+a1x+a2x2p(a; x) = a_0 + a_1 x + a_2 x^2, and q(b;x)=b0+b1x+b2x2q(b; x) = b_0 + b_1x + b_2 x^2, then the system p(a;x)p(a; x), q(b;x)q(b; x) has a simeltaneous zero iff:
[a2a1a000a2a1a0b2b1b000b2b1b0][1xx2x3]=0Ax=0 \begin{aligned} &\begin{bmatrix} a_2 & a_1 & a_0 & 0 \\ 0 & a_2 & a_1 & a_0 \\ b_2 & b_1 & b_0 & 0\\ 0 & b_2 & b_1 & b_0\\ \end{bmatrix} \begin{bmatrix} 1 \\ x \\ x^2 \\ x^3 \end{bmatrix} = 0 \\ &A x = 0 \end{aligned}

§ Big idea


The matrix is setup in such a way that any solution vector vv such that Qv=0Qv = 0 will be of the form v=(α3,α2,α,1)v = (\alpha^3, \alpha^2, \alpha, 1). That is, the solution vector is a polynomial , such that Qv=0Qv = 0. Since Qv=0Qv = 0, we have that a2α2+a1α+a0=0a_2 \alpha^2 + a_1 \alpha + a_0 = 0, and b2α2+b1α+b0=0b_2 \alpha^2 + b_1 \alpha + b_0 = 0.

§ Proof




a2w+a1x+a0y=0b2w+a1x+a0y=0 a_2 w + a_1 x + a_0 y = 0 \\ b_2 w + a_1 x + a_0 y = 0 \\

Similarly:
a2x+a1y+a0z=0b2x+a1y+a0z=0 a_2 x + a_1 y + a_0 z = 0 \\ b_2 x + a_1 y + a_0 z = 0 \\

Since the coefficients of the two systems are the same, we must have that (w,x,y)(w, x, y) and (x,y,z)(x, y, z) are linearly dependent. That is:
(w,x,y)=α(x,y,z)w=αx=α2y=α3z (w, x, y) = \alpha (x, y, z) \\ w = \alpha x = \alpha^2 y = \alpha^3 z \\

We can take z=1z = 1 arbitrarily, giving us a vector of the form (w,x,y,z)=(α3,α2,α,1)(w, x, y, z) = (\alpha^3, \alpha^2, \alpha, 1), which is the structure of the solution we are looking for!

§ References