You aren't going to get a response that long. That is just the limitations of LLM's. If you do manage to get something that long it won't make sense as it can't hold enough context as it generates.
I've won NaNoWriMo twice and I can confirm that writing your own does not necessarily result in a cool or coherent story. One of the two is likely better than an LLM could come up with, though.
The limited context lengths for local LLMs will be a barrier to write 10k words in a single prompt. Approaches to this is to have the LLM have a conversation with itself or other LLMs. There are prompts out there that can simulate this, but you will need to intervene every few hundred words or so. Check out ‘AutoGen’ frameworks that can orchestrate this for you. CrewAI is one of the better ones. hope this helps
Ollama provides a Python API which may be useful. You could have it generate the story in chunks, having it generate a list of key points which get passed to subsequent prompts. Maybe...