Rudimentary self-hosted webapp to generate short stories for children with a self-hosted LLM.
Just for playing around and for demonstration purposes.
Currently it's using ollama hosting mistral-openorca. The model is configurable.
Without CUDA:
docker compose up
With CUDA:
docker compose -f docker-compose.cuda.yml up