Venice Ai
Complete Venice AI platform — text generation, vision/image analysis, web search, X/Twitter search, embeddings, TTS, speech-to-text, image generation, backgr...
Complete Venice AI platform — text generation, vision/image analysis, web search, X/Twitter search, embeddings, TTS, speech-to-text, image generation, backgr...
Real data. Real impact.
Emerging
Developers
Per week
Open source
Skills give you superpowers. Install in 30 seconds.
Venice AI is the AI platform for people who don't want Big Tech watching over their shoulder. Unlike OpenAI, Anthropic, and Google — where every prompt is logged, analyzed, and potentially used to train future models — Venice offers true privacy with zero data retention on private models. Your conversations stay yours.
Venice is also uncensored: no content filters, no refusals, no "I can't help with that." The AI answers what you actually ask, whether that's creative fiction, security research, controversial topics, or anything else mainstream AI platforms won't touch.
Why agents should care:
This skill gives you the complete Venice platform in one place.
⚠️ API changes: If something doesn't work as expected, check docs.venice.ai — the API specs may have been updated since this skill was written.
vn_...)Option A: Environment variable
export VENICE_API_KEY="vn_your_key_here"
Option B: Clawdbot config (recommended)
// ~/.clawdbot/clawdbot.json { skills: { entries: { "venice-ai": { env: { VENICE_API_KEY: "vn_your_key_here" } } } } }
python3 {baseDir}/scripts/venice.py models --type text
| Script | Purpose |
|---|---|
| Text generation, vision analysis, models, embeddings, TTS, transcription |
| Image generation, background removal |
| Video generation (Sora, WAN, Runway) |
| Music generation (queue-based async) |
| Image upscaling |
| AI image editing, multi-image editing |
Venice has a huge model catalog spanning text, image, video, audio, and embeddings.
# List all text models python3 {baseDir}/scripts/venice.py models --type textList image models
python3 {baseDir}/scripts/venice.py models --type image
List all model types
python3 {baseDir}/scripts/venice.py models --type text,image,video,audio,embedding
Get details on a specific model
python3 {baseDir}/scripts/venice.py models --filter grok
| Need | Recommended Model | Why |
|---|---|---|
| Cheapest text | | Tiny, fast, efficient |
| Best uncensored | | Venice's own uncensored model |
| Best private + smart | | Great reasoning, efficient |
| Vision/multimodal | | Analyze images, video, audio |
| Best coding | | Massive coder model |
| Frontier fast | | Fast, 262K context |
| X/Twitter search | or | Grok models + |
| Frontier max quality | | Best overall quality |
| Reasoning | | Strong chain-of-thought |
| Web search | Any model + | Built-in web search |
# Simple prompt python3 {baseDir}/scripts/venice.py chat "What is the meaning of life?"Choose a model
python3 {baseDir}/scripts/venice.py chat "Explain quantum computing" --model deepseek-v3.2
System prompt
python3 {baseDir}/scripts/venice.py chat "Review this code" --system "You are a senior engineer."
Read from stdin
echo "Summarize this" | python3 {baseDir}/scripts/venice.py chat --model qwen3-4b
Stream output
python3 {baseDir}/scripts/venice.py chat "Write a story" --stream
# Auto web search (model decides when to search) python3 {baseDir}/scripts/venice.py chat "What happened in tech news today?" --web-search autoForce web search with citations
python3 {baseDir}/scripts/venice.py chat "Current Bitcoin price" --web-search on --web-citations
Web scraping (extracts content from URLs in prompt)
python3 {baseDir}/scripts/venice.py chat "Summarize: https://example.com/article" --web-scrape
Use Grok models to search X (Twitter) for real-time posts and discussions:
# Search X for latest AI news python3 {baseDir}/scripts/venice.py chat "latest AI news from X today" \ --model grok-41-fast --x-searchCombine X search with web search
python3 {baseDir}/scripts/venice.py chat "What are people saying about OpenAI?"
--model grok-4-20-beta --x-search --web-search auto
Note:
--x-search only works with Grok models (grok-*). It sets enable_x_search: true in venice_parameters, which routes search through xAI's infrastructure.
# Use Venice's own uncensored model python3 {baseDir}/scripts/venice.py chat "Your question" --model venice-uncensoredDisable Venice system prompts for raw model output
python3 {baseDir}/scripts/venice.py chat "Your prompt" --no-venice-system-prompt
Venice supports extended reasoning/thinking modes with fine-grained effort control:
# Use a reasoning model with effort control python3 {baseDir}/scripts/venice.py chat "Solve this math problem..." \ --model kimi-k2-5 --reasoning-effort highMinimal reasoning (faster, cheaper)
python3 {baseDir}/scripts/venice.py chat "Simple question"
--model qwen3-4b --reasoning-effort minimalMaximum reasoning (slowest, most thorough)
python3 {baseDir}/scripts/venice.py chat "Complex analysis"
--model claude-opus-4-6 --reasoning-effort maxStrip thinking from output (result only)
python3 {baseDir}/scripts/venice.py chat "Debug this code" --model qwen3-4b --strip-thinking
Disable reasoning entirely
python3 {baseDir}/scripts/venice.py chat "Quick answer" --model qwen3-4b --disable-thinking
Reasoning effort values (not all models support all levels):
| Value | Description |
|---|---|
| No reasoning (fastest) |
| Very brief thinking |
| Light reasoning |
| Balanced (often default) |
| Thorough reasoning |
| Extended reasoning |
| Maximum reasoning budget |
For maximum privacy, use End-to-End Encryption with supported models:
# Enable E2EE (prompts encrypted client-side before reaching Venice) python3 {baseDir}/scripts/venice.py chat "sensitive analysis" \ --model some-e2ee-model --enable-e2ee
Privacy tiers on Venice:
# Temperature and token control python3 {baseDir}/scripts/venice.py chat "Be creative" --temperature 1.2 --max-tokens 4000JSON output mode
python3 {baseDir}/scripts/venice.py chat "List 5 colors as JSON" --json
Prompt caching (for repeated context — up to 90% cost savings)
python3 {baseDir}/scripts/venice.py chat "Question" --cache-key my-session-123
Show usage stats and balance
python3 {baseDir}/scripts/venice.py chat "Hello" --show-usage
Use a Venice character
python3 {baseDir}/scripts/venice.py chat "Tell me about yourself" --character venice-default
Analyze images using multimodal vision models. Supports local files, URLs, and data URLs.
# Analyze a local image python3 {baseDir}/scripts/venice.py analyze photo.jpg "What's in this image?"Analyze with default prompt (describe in detail)
python3 {baseDir}/scripts/venice.py analyze photo.jpg
Analyze from URL
python3 {baseDir}/scripts/venice.py analyze "https://example.com/image.jpg" "Describe the scene"
Choose vision model
python3 {baseDir}/scripts/venice.py analyze diagram.png "Explain this diagram"
--model qwen3-vl-235b-a22bStream the analysis
python3 {baseDir}/scripts/venice.py analyze photo.jpg "Identify all objects" --stream
Count tokens used
python3 {baseDir}/scripts/venice.py analyze photo.jpg "Analyze this" --show-usage
Vision-capable models:
qwen3-vl-235b-a22b, claude-opus-4-6, gpt-5.2, and others — check --list-models for current availability.
Supported image formats: JPEG, PNG, WebP, GIF, BMP
Generate vector embeddings for semantic search, RAG, and recommendations:
# Single text python3 {baseDir}/scripts/venice.py embed "Venice is a private AI platform"Multiple texts (batch)
python3 {baseDir}/scripts/venice.py embed "first text" "second text" "third text"
From file (one text per line)
python3 {baseDir}/scripts/venice.py embed --file texts.txt
Output as JSON
python3 {baseDir}/scripts/venice.py embed "some text" --output json
Model:
text-embedding-bge-m3 (private, $0.15/M tokens)
Convert text to speech with 60+ multilingual voices:
# Default voice python3 {baseDir}/scripts/venice.py tts "Hello, welcome to Venice AI"Choose a voice
python3 {baseDir}/scripts/venice.py tts "Exciting news!" --voice af_nova
List available voices
python3 {baseDir}/scripts/venice.py tts --list-voices
Custom output path
python3 {baseDir}/scripts/venice.py tts "Some text" --output /tmp/speech.mp3
Adjust speed
python3 {baseDir}/scripts/venice.py tts "Speaking slowly" --speed 0.8
Popular voices:
af_sky, af_nova, am_liam, bf_emma, zf_xiaobei (Chinese), jm_kumo (Japanese)
Model:
tts-kokoro (private, $3.50/M characters)
Transcribe audio files to text:
# Transcribe a file python3 {baseDir}/scripts/venice.py transcribe audio.wavWith timestamps
python3 {baseDir}/scripts/venice.py transcribe recording.mp3 --timestamps
From URL
python3 {baseDir}/scripts/venice.py transcribe --url https://example.com/audio.wav
Supported formats: WAV, FLAC, MP3, M4A, AAC, MP4
Model:
nvidia/parakeet-tdt-0.6b-v3 (private, $0.0001/audio second)
python3 {baseDir}/scripts/venice.py balance
| Feature | Cost |
|---|---|
| Image generation | ~$0.01-0.03 per image |
| Background removal | ~$0.02 |
| Image upscale | ~$0.02-0.04 |
| Image edit (single) | ~$0.04 |
| Image multi-edit | ~$0.04-0.08 |
| Video (WAN) | ~$0.10-0.50 |
| Video (Sora) | ~$0.50-2.00 |
| Video (Runway) | ~$0.20-1.00 |
| Music generation | varies by model/duration |
Use
--quote with video/music commands to check pricing before generation.
# Basic generation python3 {baseDir}/scripts/venice-image.py --prompt "a serene canal in Venice at sunset"Multiple images
python3 {baseDir}/scripts/venice-image.py --prompt "cyberpunk city" --count 4
Custom dimensions
python3 {baseDir}/scripts/venice-image.py --prompt "portrait" --width 768 --height 1024
List available models and styles
python3 {baseDir}/scripts/venice-image.py --list-models python3 {baseDir}/scripts/venice-image.py --list-styles
Use specific model and style
python3 {baseDir}/scripts/venice-image.py --prompt "fantasy" --model flux-2-pro
--style-preset "Cinematic"Reproducible results with seed
python3 {baseDir}/scripts/venice-image.py --prompt "abstract" --seed 12345
PNG format with no watermark
python3 {baseDir}/scripts/venice-image.py --prompt "product shot"
--format png --hide-watermark
Key flags:
--prompt, --model (default: flux-2-max), --count, --width, --height, --format (webp/png/jpeg), --resolution (1K/2K/4K), --aspect-ratio, --negative-prompt, --style-preset, --cfg-scale (0-20), --seed, --safe-mode, --hide-watermark, --embed-exif, --steps
Remove the background from any image, producing a transparent PNG:
# From local file python3 {baseDir}/scripts/venice-image.py --background-remove photo.jpgSpecify output path
python3 {baseDir}/scripts/venice-image.py --background-remove photo.jpg --output cutout.png
From URL
python3 {baseDir}/scripts/venice-image.py --background-remove
"https://example.com/product.jpg" --output product-transparent.png
The output is always a PNG with a transparent background (alpha channel). Works best with clear subject/background separation.
# 2x upscale python3 {baseDir}/scripts/venice-upscale.py photo.jpg --scale 24x with AI enhancement
python3 {baseDir}/scripts/venice-upscale.py photo.jpg --scale 4 --enhance
Enhanced with custom prompt
python3 {baseDir}/scripts/venice-upscale.py photo.jpg --enhance --enhance-prompt "sharpen details"
From URL
python3 {baseDir}/scripts/venice-upscale.py --url "https://example.com/image.jpg" --scale 2
Key flags:
--scale (1-4, default: 2), --enhance (AI enhancement), --enhance-prompt, --enhance-creativity (0.0-1.0), --url, --output
AI-powered editing where the model interprets your prompt to modify the image:
# Add elements python3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt "add sunglasses"Modify scene
python3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt "change the sky to sunset"
Remove objects
python3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt "remove the person in background"
From URL
python3 {baseDir}/scripts/venice-edit.py --url "https://example.com/image.jpg"
--prompt "colorize this black and white photo"Specify output location
python3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt "add snow" --output result.png
Compose or blend multiple images together using advanced edit models:
# Combine 2 images python3 {baseDir}/scripts/venice-edit.py --multi-edit base.jpg overlay.png \ --prompt "merge these images seamlessly"Layer 3 images with model selection
python3 {baseDir}/scripts/venice-edit.py --multi-edit bg.jpg subject.png detail.png
--prompt "compose these layers into one image" --model flux-2-max-editMix local file and URL
python3 {baseDir}/scripts/venice-edit.py --multi-edit local.jpg "https://example.com/img.png"
--prompt "blend these two photos" --output blended.png
Multi-edit models:
flux-2-max-edit (default), qwen-edit, gpt-image-1-5-edit
Note: The standard edit endpoint uses Qwen-Image which has some content restrictions. Multi-edit uses Flux-based models.
# Get price quote first python3 {baseDir}/scripts/venice-video.py --quote --model wan-2.6-image-to-video --duration 10sImage-to-video (WAN - default)
python3 {baseDir}/scripts/venice-video.py --image photo.jpg --prompt "camera pans slowly"
--duration 10sImage-to-video (Sora)
python3 {baseDir}/scripts/venice-video.py --image photo.jpg --prompt "cinematic"
--model sora-2-image-to-video --duration 8s --aspect-ratio 16:9 --skip-audio-paramVideo-to-video (Runway Gen4)
python3 {baseDir}/scripts/venice-video.py --video input.mp4 --prompt "anime style"
--model runway-gen4-turbo-v2vList models with available durations
python3 {baseDir}/scripts/venice-video.py --list-models
Key flags:
--image or --video, --prompt, --model (default: wan-2.6-image-to-video), --duration, --resolution (480p/720p/1080p), --aspect-ratio, --audio/--no-audio, --quote, --timeout
Models:
--aspect-ratio, use --skip-audio-paramVenice supports AI music generation via a queue-based async API (similar to video generation). Music is generated server-side and polled for completion.
# Get price quote first python3 {baseDir}/scripts/venice-music.py --quote --model elevenlabs-music --duration 60Generate instrumental music
python3 {baseDir}/scripts/venice-music.py --prompt "epic orchestral battle theme" --instrumental
Generate music with lyrics
python3 {baseDir}/scripts/venice-music.py
--prompt "upbeat pop summer song"
--lyrics "Verse 1: Walking down the beach / feeling the heat..."Control duration
python3 {baseDir}/scripts/venice-music.py --prompt "ambient piano meditation" --duration 30
Specify output location
python3 {baseDir}/scripts/venice-music.py --prompt "jazz café background"
--output ~/Music/venice-jazz.mp3List available audio models
python3 {baseDir}/scripts/venice-music.py --list-models
Don't delete from server after download (useful for re-downloading)
python3 {baseDir}/scripts/venice-music.py --prompt "..." --no-delete
Clean up server-side media after downloading with --no-delete
python3 {baseDir}/scripts/venice-music.py --complete QUEUE_ID
Parameters:
| Flag | Description |
|---|---|
| Music description (style, mood, genre, instruments) |
| Model ID (default: ) |
| Duration in seconds |
| Optional lyrics text for vocal generation |
| Force instrumental (no vocals) |
| Voice selection for vocal tracks |
| Language code (e.g., , , ) |
| Get price estimate without generating |
| Max wait time in seconds (default: 300) |
| Status check interval (default: 10s) |
Prompt tips for music:
"chill lo-fi hip hop with piano and rain ambiance, 85 BPM""cinematic orchestral swell, strings and brass, dramatic tension""acoustic folk guitar, warm and intimate, fingerpicking style"Use
--web-search on --web-citations to build a research workflow. Venice searches the web, synthesizes results, and cites sources — all in one API call.
With Grok models and
--x-search, you get real-time access to X posts and discussions. Great for trend monitoring, social listening, and news research.
Venice's uncensored models work for both text AND images. No guardrails blocking legitimate creative use cases.
When processing sensitive data:
--enable-e2ee for encrypted prompt deliveryIf you're running an agent loop that sends the same system prompt repeatedly, use
--cache-key to get up to 90% cost savings.
# Analyze → describe → generate matching image python3 scripts/venice.py analyze original.jpg "describe the style and composition" > desc.txt python3 scripts/venice-image.py --prompt "$(cat desc.txt)" --model flux-2-max
Combine TTS and transcription: generate spoken content with
tts, process audio with transcribe. Both are private inference.
--quote to estimate video cost# Generate background music, then use it for video python3 scripts/venice-music.py --prompt "cinematic adventure theme" --output bgm.mp3 python3 scripts/venice-video.py --image scene.jpg --prompt "epic journey" --audio-url bgm.mp3
# Generate → remove background → use in video python3 scripts/venice-image.py --prompt "product on white background" --format png python3 scripts/venice-image.py --background-remove output.png --output product-clean.png
| Problem | Solution |
|---|---|
| Set env var or configure in |
| Verify at venice.ai/settings/api |
| Run to see available; use for new models |
| Rate limited | Check output |
| Video stuck | Videos can take 1-5 min; use for long ones |
| Vision not working | Ensure you're using a vision-capable model (e.g., ) |
no effect | Only works with Grok models () |
| Music timeout | Music can take 2-5 min; increase |
| Background removal quality | Works best with clear subject/background contrast |
No automatic installation available. Please visit the source repository for installation instructions.
View Installation Instructions1,500+ AI skills, agents & workflows. Install in 30 seconds. Part of the Torly.ai family.
© 2026 Torly.ai. All rights reserved.