They are related, and we frequently assume they are close enough that it doesn’t matter, but they are different.
They did it with a custom language model. I really want to give this a try with llama2 embeddings but haven't had the bandwidth yet (and llama2's embedding vectors are inconveniently huge, but that's a different problem).
Are there any good sources to learn more about that?