AI Skill

What is the LuvVoice skill?

The LuvVoice skill gives an AI agent a lightweight, file-based path to text-to-speech. Instead of connecting to a server first, the agent reads one markdown file, follows the instructions inside it, and calls the API directly.

Download ZIP

Install one file

Drop the skill into the right client folder and the agent can discover the workflow immediately.

Teach the agent

The markdown file explains endpoints, parameters, defaults, and how to recover from common issues.

Return audio fast

Once loaded, the agent can generate speech, pick voices, and hand back a usable audio result.

How the skill workflow fits together

A skill is still a real integration surface, just a lighter one. The agent gets explicit instructions, then uses them to execute the task without you needing to run extra infrastructure.

1. User asks for speech

The request can be natural language or an explicit command, depending on the client.

2. Agent reads SKILL.md

The file defines prerequisites, endpoint usage, default voice behavior, and response handling.

3. Audio comes back

The agent makes the API call, receives the payload, and returns a playable link or file.

Installation stays client-specific, not workflow-specific

The workflow does not change across clients. What changes is only the file path and how that client chooses to invoke the skill.

Download the ready-made package

If you prefer not to curl the raw file directly, download the ZIP, extract it, and place the skill in the directory shown below.

Download ZIP
Install location
Factory Droid
~/.factory/skills/luvvoice-tts/SKILL.md
Invocation
Factory Droid
/luvvoice-tts or auto-detected by Droid
bash - set the API token
export LUVVOICE_API_TOKEN=your_token_here
bash - download for Factory Droid
mkdir -p ~/.factory/skills/luvvoice-tts
curl -o ~/.factory/skills/luvvoice-tts/SKILL.md \
  https://raw.githubusercontent.com/luvvoice/luvvoice-mcp/main/.factory/skills/luvvoice-tts/SKILL.md

What the skill handles once installed

The point of the skill is not just setup convenience. It also gives the agent a reusable operating model for synthesis, discovery, and sensible defaults.

Text-to-Speech

Convert arbitrary text to MP3 output with voice, speed, pitch, and volume controls.

Voice Discovery

Filter the catalog by language or gender, then keep a shortlist of reliable starting voices.

Smart Defaults

When the user is vague, the agent can infer language, pick a natural default, and split long text safely.

Popular starting voices

The full voice catalog is still available at runtime. These are simply the voices people reach for first when testing the skill.

200+ voices available via the API
Jenny
voice-001
English (US)
Xiaoxiao
voice-093
Chinese (Mandarin)
Nanami
voice-120
Japanese
Alvaro
voice-050
Spanish
Katja
voice-111
German
Denise
voice-140
French

Keep the skill when you want lighter setup

The decision is less about features and more about integration posture. Both paths reach the same LuvVoice API, but they do it in different ways.

Choose the skill

You want the fastest path from zero to working speech output.
A single markdown file is easier to distribute than a tool server.
Direct API calls are enough for the workflow you need.

Choose MCP

You want structured tools instead of instruction-driven API calls.
Your client already has a strong MCP workflow built in.
You may later need richer transports or shared deployments.

Ready to install it?

The practical sequence is still short: create a token, place the file, and let the agent follow the instructions from there.

Give your coding agent a voice in under a minute

Start with the skill if you want the lightest install path, and keep the MCP route nearby if your workflow later grows into a tool-based integration.