It can do that for some images, and that's a training mistake.
There's no way it can do that for more than a tiny fraction.
"we bias our search towards duplicated training examples
because these are orders of magnitude more likely to be
memorized than non-duplicated example"