I’ve tried their 1.1B model. The only hiccup was that it seems to require mlx 0.10.0 which is what’s in requirements.txt. You also have to place the llama tokenizer file into the model dir - they do not distribute it. The models published for MLX do not seem to be instruction tuned, so with their default prompt they get repetitive. But I suppose you could convert the instruction tuned checkpoints with the script in the repo.