That is literally how their search ranking team works. Here is a search strategy, here is another. Tell me which is better based on a bunch of random queries.
For spammy websites, you don’t review every one individually. But you have an algorithm that e.g. downranks them, and you AB test it.
They probably already have this strategy now, just not deployed for one reason or another. (For example, it might be expensive and unreliable to scan every website.) But if Google loses enough traffic to a competitor, they’ll be forced to do something.