Recently I've been reading a draft version of a book by Jiří Matoušek. The book seems to take problems that are combinatorial or algorithmic in nature and solves them using ideas of linear algebra. What's amazing about the problems is that on first glance they don't seem to have anything to do with linear algebra at all. But the book solves them using linear algebra anyway: it forms certain matrices, multiplies them, takes their determinants, argues about their ranks and somehow produces a solution.
In this note, I want to share one of my favorite problems from that book. You do need to know a little bit of linear algebra to understand what's going on in the solution. But it's not anything highly technical. In fact, this solution, like most of the other solutions in the book, hinges on just one fact: the rank of the product of matrices is no larger than the minimum of the ranks of the factors. Even if you don't know why this is true, you can still follow along. But if you don't know what the rank of a matrix is, things are probably going to be hard for you to follow.
I think I've done enough talking. It's time to see the problem.
There can't exist four points on the plane such that the distance between any pair of them is an odd integer.
I'm going to be honest here. This is the sort of problem I wouldn't know how to even begin to solve. I guess I would start by drawing some pictures and noticing that the statement seems to hold true no matter what. But I would then eventually give up without making any sort of significant progress.
But this article isn't about my problem-solving skills or the lack thereof. This is about how we're going to solve a problem with no apparent connection to linear algebra with linear algebra. So let's begin!
For the sake of contradiction, we're going to assume that there are four such points that the distance between any pair of them is odd. Let the points be , , and . Here we are treating the points as vectors in . Without any loss of generality, we took the first vector to be the null vector. By assumption, all of , , , , and are odd.
Let be the very familiar dot product of and . Here is the angle between the vectors and .
Since the square of an odd number leaves a remainder of when divided by , it follows that . Also from the cosine theorem, we have .
Now consider the matrix
Note that,
Taking the determinant, we get . In particular, we have and so .
This is the time we need to know about the rank of a matrix. If you aren't familiar with it, just think of it as the maximum number of linearly independent rows of a matrix. Some rows of a matrix are said to be linearly independent if none of the rows can be expressed as a linear combination of the other rows. If none of these terms seem familiar, look them up. You don't need to know much linear algebra to know what these terms mean.
Since , it means that the rows of are linearly independent. So the maximum number of linearly independent rows of is simply equal to the number of rows of . In other words, .
The crucial part of the proof is to notice that where
Here , etc. Since has two rows, it means that the rank of is at most two.
Now we're ready to use the fact mentioned at the beginning of the article: the rank of the product of matrices is no larger than the minimum of the ranks of the factors. So the rank of can't be larger than the rank of which itself is at most . So, . But that's a contradiction since we already established that .
So, we're wrong in assuming the existence of the points. This concludes the proof.
If you haven't seen a proof like this before, this should all feel like magic. It takes a while (at least it did for me) to really understand what's going on. But damn is it elegant!
Easy Math Editor
This discussion board is a place to discuss our Daily Challenges and the math and science related to those challenges. Explanations are more than just a solution — they should explain the steps and thinking strategies that you used to obtain the solution. Comments should further the discussion of math and science.
When posting on Brilliant:
*italics*
or_italics_
**bold**
or__bold__
paragraph 1
paragraph 2
[example link](https://brilliant.org)
> This is a quote
\(
...\)
or\[
...\]
to ensure proper formatting.2 \times 3
2^{34}
a_{i-1}
\frac{2}{3}
\sqrt{2}
\sum_{i=1}^3
\sin \theta
\boxed{123}
Comments
I can't remember how long it's been since I last wrote something like this on Brilliant. It's definitely been while and it was a lot of fun to do. Please let me know if you want to see more stuff like this. As always, any sort of feedback is very much appreciated.
Log in to reply
This was indeed nice. I definitely enjoying reading your notes.
I do get the idea of the proof. But, to me, it is feeling like a magic still. How do we get an intuition that this method will actually do work? Is there any deep connection lying behind always?
Which book is this?
Log in to reply
Thirty-three Miniatures: Mathematical and Algorithmic Applications of Linear Algebra