To be clear (apologies if I haven't made it so) this is not an LLM. This is an implementation of Rasa leveraging Langchain under the hood.
A user technically does not need to dig into Langchain themselves, but they would want to if they find their query results sub-optimal.
There are a many indexing strategies and superficial parameters you could modify to tune output response. They are mentioned in the README.md.