Aside from the artwork and the laugh track you'll hear, everything else is generative, including: dialogue, speech, direction (camera cuts, character focus, shot length, scene length, etc), character movement, and music.
Nothing, Forever is built using a combination of machine learning, generative algorithms (we use 'generative' here in a non-academic sense), and cloud services. Our stack is mostly comprised of Python + TensorFlow for our ML models, TypeScript + Azure Functions and Heroku for our backend, and C# + Unity for the client, with some neural voice APIs thrown into the mix.
Heading into the future, we plan to leverage OpenAI's davinci models for our dialogue -- we actually have an integration already, but the cost is too prohibitive to run full time -- as well as leveraging Stable Diffusion for art generation.
We're launching an experimental media lab, https://www.mismatchmedia.com/ along with the show, for follow-up projects in the future.
Feel free to reach out to me personally if you have any questions or want to discuss more, skyler at mismatchmedia dotcom.