I assume quantized models will run a lot better. TheBloke already seems like he's on it.
https://huggingface.co/TheBloke/CodeLlama-13B-fp16
Because codellama is llama based it may just work possibly?