podcli takes a long-form podcast and turns it into a complete content operation:
Record episode
↓
Transcribe (Whisper, speaker detection)
↓
Find viral moments (Claude AI + audio energy + knowledge base)
↓
Render clips (9:16, captions, smart crop, normalized audio)
↓
Generate content package (titles, descriptions, thumbnails, SEO) ← PodStack
↓
Publish with optimization checklist ← PodStack
↓
Review performance ← PodStackThe first half is video processing — podcli's core engine. The second half is content workflow — powered by PodStack, a set of Claude Code slash commands that ship with podcli. Both halves are deeply integrated: the clip suggestion engine reads from your PodStack knowledge base, uses your title formulas and voice rules, checks the episode database for duplicates, and outputs MCP-aligned fields that flow through to export.
./setup.sh --ui
# → http://localhost:3847Drag your video into the Web UI, or use the CLI:
./podcli process episode.mp4 --transcript transcript.txt --top 8podcli uses Claude to analyze your transcript against your show's knowledge base, finding the most viral moments. It scores each one on 4 dimensions, suggests clips with multi-cut segments (cutting out filler), and lets you toggle them on/off before rendering.
Clips come out as upload-ready Shorts: 1080x1920, 9:16 vertical, with burned-in captions, normalized audio, and your logo.
Open the project in Claude Code and run:
/prep-episodeThis runs the PodStack pipeline — a gstack-style workflow that gives you:
Run /publish-checklist when uploading. A week later, run /retro-episode with your YouTube Studio stats to see what worked and what to improve.
| Video Engine (podcli core) | Content Workflow (PodStack) | |
|---|---|---|
| What | Transcription, clip detection, rendering | Titles, descriptions, thumbnails, publishing |
| How | Python + FFmpeg + Whisper + OpenCV + Claude/Codex | Claude Code slash commands |
| Interface | Web UI, CLI, MCP tools | /slash-commands in Claude Code |
| Output | .mp4 files ready to upload | Content packages ready to paste into YouTube |
Both halves share the same knowledge base (.podcli/knowledge/) — your show's brand, voice, title formulas, episode database, and style guide. Set it up once, everything stays on-brand.
Speaker (MM:SS), JSON, drag-drop .txt / .srt / .vtt/process-transcript — extract and score best moments from any transcript/generate-titles — 8 titles per clip with 6-point verification checklist/generate-descriptions — descriptions + hashtags + SEO keywords/plan-thumbnails — thumbnail text + designer briefs for both formats/review-content — paranoid brand check (banned words, voice, title rules)/prep-episode — full pipeline: transcript → publish-ready package/publish-checklist — pre/post-publish optimization/retro-episode — performance analysis after publishing.md files that teach the AI your brand, voice, and stylelocalhost:3847./podcli process video.mp4 --top 5| Tool | Install |
|---|---|
| Node.js >= 18 | nodejs.org |
| Python >= 3.10 | python.org |
| FFmpeg | brew install ffmpeg / sudo apt install ffmpeg |
| Claude Code (optional) | docs.anthropic.com — needed for PodStack slash commands |
| Codex (optional) | openai.com/codex — alternative AI engine for clip suggestion (auto-detected if Claude is unavailable) |
git clone https://github.com/nmbrthirteen/podcli.git
cd podcli
chmod +x setup.sh podcli
./setup.shThis will:
.podcli/ data directory./setup.sh --ui
# → http://localhost:3847.txt file, paste Speaker (MM:SS) text, or auto-transcribe with Whisper# Auto-transcribe + suggest top 5 clips + export
./podcli process video.mp4
# With existing transcript
./podcli process video.mp4 --transcript transcript.txt --top 5
# Full options
./podcli process video.mp4 \
--transcript transcript.txt \
--top 8 \
--caption-style branded \
--crop center \
--logo logo.png./podcli presets save myshow --caption-style branded --logo logo.png --top 5
./podcli presets list
./podcli process video.mp4 --preset myshowOpen the project in Claude Code, then use slash commands:
# Full pipeline — transcript to publish-ready package
/prep-episode
# Individual steps
/process-transcript # extract moments from a transcript
/generate-titles # get 8 title options for a clip
/generate-descriptions # get descriptions + hashtags
/plan-thumbnails # get thumbnail briefs for your designer
/review-content # brand and quality review
/publish-checklist # pre/post-publish ops
/retro-episode # performance analysisOr just paste a transcript — Claude auto-detects the input and runs the right command.
The knowledge base is what makes podcli understand your show. Drop .md files into .podcli/knowledge/ and both the video engine and content workflow use them. The clip suggestion engine reads 8 of these files (prioritized by relevance), checks the episode database for duplicate avoidance, and applies your voice rules and title formulas when generating suggestions.
PodStack ships with 13 starter templates that you fill in with your show's details:
| File | What It Teaches The AI |
|---|---|
00-master-instructions.md | Auto-detection rules, decision tree, quality gates |
01-brand-identity.md | Show name, positioning, tagline, hosts, format |
02-voice-and-tone.md | Voice fingerprint, banned words, the Coffee Test |
03-episodes-database.md | Episode tracking, existing shorts (for dedup) |
04-shorts-creation-guide.md | Moment types, selection criteria, extraction process |
05-title-formulas.md | Title shapes, rules, templates by content type |
06-descriptions-template.md | Description formulas, hashtag library, SEO keywords |
07-thumbnail-guide.md | Layouts, brand colors, typography, visual specs |
08-topics-themes.md | Core topics, cross-cutting themes, audience map |
09-content-workflow.md | End-to-end workflow phases, handoff specs |
10-internal-processing.md | Auto-execution rules, internal quality gates |
11-inspiration-channels.md | Reference channels, viral hooks, hybrid formulas |
12-quick-reference.md | Copy-paste hooks, hashtags, CTAs, checklists |
Manage via the web UI at /knowledge.html (drag & drop, inline editor) or through the knowledge_base MCP tool.
podcli is a Model Context Protocol server — Claude can use it as a tool to create clips through conversation.
Claude Desktop — add to claude_desktop_config.json:
{
"mcpServers": {
"podcli": {
"command": "node",
"args": ["/path/to/podcli/dist/index.js"],
"env": {
"PYTHON_PATH": "/path/to/podcli/venv/bin/python3"
}
}
}
}Claude Code:
claude mcp add podcli -- node /path/to/podcli/dist/index.jsRun ./setup.sh --mcp to get the exact config with your paths filled in.
| Tool | Description |
|---|---|
transcribe_podcast | Transcribe audio/video with Whisper + speaker detection |
suggest_clips | Submit clip suggestions (includes duplicate check) |
create_clip | Render a single short-form clip as a vertical short |
batch_create_clips | Render multiple clips in one batch |
knowledge_base | Read/manage podcast context files (hosts, style, audience, etc.) |
manage_assets | Register/list reusable assets (logos, videos) |
clip_history | View previously created clips, check for duplicates |
get_ui_state | Read current session state and get workflow next-step guidance |
modify_clip | Adjust a suggested clip's timing, title, or caption style (or delete it) |
toggle_clip | Select or deselect a suggested clip for export |
update_settings | Update rendering settings (caption style, crop strategy, logo, outro) |
list_outputs | List all rendered clip files in the output directory |
manage_presets | Save, load, list, or delete rendering presets |
analyze_energy | Analyze audio energy levels to find high-energy moments |
set_video | Set the working video file without transcribing |
import_transcript | Import an external transcript with word-level timestamps (skips Whisper) |
parse_transcript | Parse raw speaker-labeled plain text into word-level timestamps |
| Style | Look |
|---|---|
| branded | Large bold text, dark box highlight on active word, gradient overlay, optional logo |
| hormozi | Bold uppercase pop-on text, yellow active word (Alex Hormozi style) |
| karaoke | Full sentence visible, words highlight progressively |
| subtle | Clean minimal white text at bottom |
podcli/
├── podcli # CLI entry point
├── setup.sh # one-command install & launch
├── package.json
├── CLAUDE.md # PodStack master config
│
├── .claude/commands/ # PodStack slash commands
│ ├── process-transcript.md
│ ├── generate-titles.md
│ ├── generate-descriptions.md
│ ├── plan-thumbnails.md
│ ├── review-content.md
│ ├── prep-episode.md
│ ├── publish-checklist.md
│ └── retro-episode.md
│
├── src/ # TypeScript
│ ├── index.ts # MCP server entry (stdio)
│ ├── server.ts # MCP tool definitions
│ ├── config/paths.ts
│ ├── models/index.ts
│ ├── handlers/ # MCP tool handlers
│ ├── services/
│ │ ├── python-executor.ts
│ │ ├── file-manager.ts
│ │ ├── asset-manager.ts
│ │ ├── clips-history.ts
│ │ ├── knowledge-base.ts
│ │ └── transcript-cache.ts
│ └── ui/
│ ├── web-server.ts # Express server + API
│ └── public/ # Frontend (React SPA)
│
├── backend/ # Python
│ ├── main.py # stdin/stdout JSON dispatcher
│ ├── cli.py # CLI entry point
│ ├── presets.py
│ ├── requirements.txt
│ ├── models/ # ML model files
│ │ └── face_detection_yunet_2023mar.onnx
│ ├── services/ # Whisper, FFmpeg, captions, face tracking, etc.
│ │ ├── face_detector.py # shared YuNet face detector
│ │ └── ...
│ └── config/
│ └── caption_styles.py
│
└── .podcli/ # local data (gitignored)
├── knowledge/ # .md context files for AI (13 templates)
├── assets/ # registered logos, videos
├── cache/transcripts/ # cached transcriptions
├── history/ # generated clip history
├── output/ # rendered clips
├── presets/ # saved configurations
└── working/ # temp filesCopy .env.example to .env (setup.sh does this automatically):
| Variable | Default | Description |
|---|---|---|
WHISPER_MODEL | base | Whisper model size (tiny, base, small, medium, large) |
WHISPER_DEVICE | auto | cpu, cuda, or auto |
PYTHON_PATH | (venv) | Path to Python binary |
PODCLI_HOME | .podcli/ | Data directory (relative to project root) |
FFMPEG_PATH | ffmpeg | Custom FFmpeg path |
LOG_LEVEL | info | Logging verbosity |
Speaker Name (00:00)
What they said goes here as plain text.
Another Speaker (00:45)
Their response text here.The time offset field (default: -1s) shifts all timestamps to sync with audio.
Content workflow powered by PodStack — inspired by gstack by Garry Tan.
MIT
nmbrthirteen/podcli
February 26, 2026
April 13, 2026
Python