Novel Mage Feature

Local Models (Ollama & LM Studio)

Run fully offline with local LLMs, or mix local and cloud.

Novel Mage supports fully local AI models through Ollama and LM Studio the two most widely used tools for running open-source LLMs on your own hardware. When you use a local model, every part of your writing workflow is offline: the AI runs on your CPU or GPU, your manuscript never leaves your machine, and there are no API costs, no token limits, and no internet dependency.

For writers, local AI is about more than privacy though that matters enormously. It's about uncapped usage. When the AI runs locally, you can generate, rewrite, summarize, and chat with no credit counter ticking, no usage tier limiting your session length, and no monthly bill for the AI component. For heavy users, this changes the economics of AI-assisted writing entirely. A local 13B model running on a consumer GPU has zero marginal cost per generation.

Novel Mage lets you mix local and cloud AI in the same project. You might use a fast local model for first-draft generation where raw speed matters, then switch to Claude for nuanced rewriting where quality is the priority all within the same novel, per task. This hybrid approach gives you the cost efficiency of local AI plus the output quality of frontier models where it counts. Local model support works with any Ollama-compatible or LM Studio-compatible model, including Llama 3, Mistral, Qwen, DeepSeek, and dozens of fine-tuned variants.

Key benefits

  • 100% offline operation your manuscript never leaves your device
  • No API costs, no token limits, no usage metering on local generation
  • Compatible with Ollama, LM Studio, and any OpenAI-compatible local endpoint
  • Works with Llama 3, Mistral, Qwen, DeepSeek, and all major open-source LLMs
  • Mix local and cloud AI per task within the same project

How it works

Install Ollama or LM Studio on your machine and pull any compatible model. In Novel Mage settings, add your local endpoint (typically localhost:11434 for Ollama or the LM Studio server port). Select the local model as your default or per-task model. All AI features generation, chat, editing tools, summaries, character interviews will call the local model. No internet required. You can add multiple local model endpoints and switch between them, or configure specific features to prefer local while others use a cloud key.

When to use it

1.

Privacy-sensitive writing manuscripts that should never be processed by a third-party server

2.

Writers who want unlimited AI generation without per-token API costs

3.

Writing in locations without reliable internet planes, trains, rural areas

4.

Authors concerned about cloud AI training on their unpublished work after the 2025 copyright cases

Related features

Try Local Models (Ollama & LM Studio) in Novel Mage

Novel Mage is $99 once lifetime access. Every feature including local models (ollama & lm studio) is included. Fully offline, local AI support, your data stays on your machine.

Free to download. No credit card required.