1Configuration

LLM Settings (Script Generation)
URL for your OpenAI-compatible server (LM Studio, Ollama, etc.)
TTS Settings (Voice Generation)
Local loads model in-process; External connects to a Gradio TTS server
URL for the Qwen3-TTS Gradio server
Language for audio synthesis only. To generate scripts in another language, edit the prompts under Prompt Settings (Advanced).
Concurrent TTS requests
Seed for reproducibility
Silence between different speakers on merge
Silence when same speaker continues on merge
Generation Settings
Size of text chunks sent to LLM (characters). Smaller = more precise, larger = more context.
Maximum tokens for LLM response. Increase if output is being truncated.
LLM Sampling Parameters
Randomness (0 = deterministic)
Nucleus sampling threshold
Top-K token filtering (0 = disabled)
Minimum probability cutoff (0 = disabled)
Penalize repeated topics
Comma-separated tokens the LLM is forbidden from generating. Use to disable thinking mode (e.g. <think>).
During script review, combine adjacent NARRATOR entries that share the same instruct into a single longer entry. Disable for better per-line voice direction control.

Prompt Customization
Instructions for the LLM on how to convert text to script. Defines output format, rules, and style guidelines.
Template for each chunk. Use {context} for chunk context and {chunk} for the text content.

Review Prompt Customization
Instructions for the LLM during the script review pass. Defines what errors to fix and how.
Template for each review batch. Use {context} for batch context and {batch} for the script entries.