I have found it pretty decent at explaining math and physics concepts, and generating some basic code. It seems over-tuned for code generation (on purpose) as sometimes it inappropriately generates code when asking non-code questions.
Overall, it performs better than GPT-3.5-turbo in many use cases. Harder to quantify the GPT-4 comparison, as there are multiple versions of GPT-4 which are rumored to have significantly varied outputs.
It's definitely worth giving a shot. The parameter size of 34B makes a big difference, and it's been found that you're still better off extremely quantizing a larger-parameter model than using unquantized smaller models.