diff --git a/docs.json b/docs.json index 6518dcbb7..7c63561c0 100644 --- a/docs.json +++ b/docs.json @@ -318,6 +318,12 @@ "tutorials/partner-nodes/bytedance/seedream-5-lite" ] }, + { + "group": "ElevenLabs", + "pages": [ + "partner-nodes/elevenlabs/elevenlabs" + ] + }, { "group": "Google", "pages": [ @@ -326,6 +332,42 @@ "tutorials/partner-nodes/google/nano-banana-2" ] }, + { + "group": "Grok", + "pages": [ + "partner-nodes/grok/grok" + ] + }, + { + "group": "HitPaw", + "pages": [ + "partner-nodes/hitpaw/hitpaw" + ] + }, + { + "group": "Magnific", + "pages": [ + "partner-nodes/magnific/magnific" + ] + }, + { + "group": "MiniMax", + "pages": [ + "partner-nodes/minimax/minimax" + ] + }, + { + "group": "PixVerse", + "pages": [ + "partner-nodes/pixverse/pixverse" + ] + }, + { + "group": "Quiver", + "pages": [ + "partner-nodes/quiver/quiver" + ] + }, { "group": "Stability AI", "pages": [ @@ -391,6 +433,12 @@ "tutorials/partner-nodes/rodin/model-generation" ] }, + { + "group": "Topaz", + "pages": [ + "partner-nodes/topaz/topaz" + ] + }, { "group": "Tripo", "pages": [ @@ -421,6 +469,18 @@ "tutorials/partner-nodes/reve/reve-image" ] }, + { + "group": "Veo", + "pages": [ + "partner-nodes/veo/veo" + ] + }, + { + "group": "Vidu", + "pages": [ + "partner-nodes/vidu/vidu" + ] + }, { "group": "Wan", "pages": [ @@ -432,6 +492,12 @@ "pages": [ "tutorials/partner-nodes/sonilo/video-to-music" ] + }, + { + "group": "WaveSpeed", + "pages": [ + "partner-nodes/wavespeed/wavespeed" + ] } ] } @@ -2414,6 +2480,12 @@ "zh/tutorials/partner-nodes/bytedance/seedream-5-lite" ] }, + { + "group": "ElevenLabs", + "pages": [ + "zh/tutorials/partner-nodes/elevenlabs/elevenlabs" + ] + }, { "group": "Google", "pages": [ @@ -2422,6 +2494,42 @@ "zh/tutorials/partner-nodes/google/nano-banana-2" ] }, + { + "group": "Grok", + "pages": [ + "zh/tutorials/partner-nodes/grok/grok" + ] + }, + { + "group": "HitPaw", + "pages": [ + "zh/tutorials/partner-nodes/hitpaw/hitpaw" + ] + }, + { + "group": "Magnific", + "pages": [ + "zh/tutorials/partner-nodes/magnific/magnific" + ] + }, + { + "group": "MiniMax", + "pages": [ + "zh/tutorials/partner-nodes/minimax/minimax" + ] + }, + { + "group": "PixVerse", + "pages": [ + "zh/tutorials/partner-nodes/pixverse/pixverse" + ] + }, + { + "group": "Quiver", + "pages": [ + "zh/tutorials/partner-nodes/quiver/quiver" + ] + }, { "group": "Stability AI", "pages": [ @@ -2487,6 +2595,12 @@ "zh/tutorials/partner-nodes/rodin/model-generation" ] }, + { + "group": "Topaz", + "pages": [ + "zh/tutorials/partner-nodes/topaz/topaz" + ] + }, { "group": "Tripo", "pages": [ @@ -2517,6 +2631,18 @@ "zh/tutorials/partner-nodes/reve/reve-image" ] }, + { + "group": "Veo", + "pages": [ + "zh/tutorials/partner-nodes/veo/veo" + ] + }, + { + "group": "Vidu", + "pages": [ + "zh/tutorials/partner-nodes/vidu/vidu" + ] + }, { "group": "Wan", "pages": [ @@ -2528,6 +2654,12 @@ "pages": [ "zh/tutorials/partner-nodes/sonilo/video-to-music" ] + }, + { + "group": "WaveSpeed", + "pages": [ + "zh/tutorials/partner-nodes/wavespeed/wavespeed" + ] } ] } @@ -4515,6 +4647,12 @@ "ja/tutorials/partner-nodes/bytedance/seedream-5-lite" ] }, + { + "group": "ElevenLabs", + "pages": [ + "ja/tutorials/partner-nodes/elevenlabs/elevenlabs" + ] + }, { "group": "Google", "pages": [ @@ -4523,6 +4661,42 @@ "ja/tutorials/partner-nodes/google/nano-banana-2" ] }, + { + "group": "Grok", + "pages": [ + "ja/tutorials/partner-nodes/grok/grok" + ] + }, + { + "group": "HitPaw", + "pages": [ + "ja/tutorials/partner-nodes/hitpaw/hitpaw" + ] + }, + { + "group": "Magnific", + "pages": [ + "ja/tutorials/partner-nodes/magnific/magnific" + ] + }, + { + "group": "MiniMax", + "pages": [ + "ja/tutorials/partner-nodes/minimax/minimax" + ] + }, + { + "group": "PixVerse", + "pages": [ + "ja/tutorials/partner-nodes/pixverse/pixverse" + ] + }, + { + "group": "Quiver", + "pages": [ + "ja/tutorials/partner-nodes/quiver/quiver" + ] + }, { "group": "Stability AI", "pages": [ @@ -4588,6 +4762,12 @@ "ja/tutorials/partner-nodes/rodin/model-generation" ] }, + { + "group": "Topaz", + "pages": [ + "ja/tutorials/partner-nodes/topaz/topaz" + ] + }, { "group": "Tripo", "pages": [ @@ -4618,6 +4798,18 @@ "ja/tutorials/partner-nodes/reve/reve-image" ] }, + { + "group": "Veo", + "pages": [ + "ja/tutorials/partner-nodes/veo/veo" + ] + }, + { + "group": "Vidu", + "pages": [ + "ja/tutorials/partner-nodes/vidu/vidu" + ] + }, { "group": "Wan", "pages": [ @@ -4629,6 +4821,12 @@ "pages": [ "ja/tutorials/partner-nodes/sonilo/video-to-music" ] + }, + { + "group": "WaveSpeed", + "pages": [ + "ja/tutorials/partner-nodes/wavespeed/wavespeed" + ] } ] } @@ -6599,4 +6797,4 @@ "destination": "/zh/:slug*" } ] -} +} \ No newline at end of file diff --git a/tutorials/image/ernie-image/ernie-image.mdx b/tutorials/image/ernie-image/ernie-image.mdx index 1192bd8a6..45f90724f 100644 --- a/tutorials/image/ernie-image/ernie-image.mdx +++ b/tutorials/image/ernie-image/ernie-image.mdx @@ -92,7 +92,7 @@ You can find all repackaged model files at [Comfy-Org/ERNIE-Image](https://huggi ### ERNIE-Image-Turbo model downloads - + Diffusion model for ERNIE-Image-Turbo. @@ -112,7 +112,7 @@ You can find all repackaged model files at [Comfy-Org/ERNIE-Image](https://huggi 📂 ComfyUI/ ├── 📂 models/ │ ├── 📂 diffusion_models/ -│ │ └── ernie-image.safetensors +│ │ └── ernie-image-turbo.safetensors │ ├── 📂 text_encoders/ │ │ ├── ministral-3-3b.safetensors │ │ └── ernie-image-prompt-enhancer.safetensors diff --git a/tutorials/partner-nodes/elevenlabs/eleven-multilingual-v2.mdx b/tutorials/partner-nodes/elevenlabs/eleven-multilingual-v2.mdx new file mode 100644 index 000000000..37c72148f --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/eleven-multilingual-v2.mdx @@ -0,0 +1,31 @@ +--- +title: "ElevenLabs Text-to-Speech — Multilingual v2 / v3 / Flash v2.5" +description: "Generate lifelike, emotionally expressive speech from text using ElevenLabs' TTS models in ComfyUI." +sidebarTitle: "Eleven Multilingual v2 / v3" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Generate speech from text using ElevenLabs' flagship Text-to-Speech models. Choose between the high-quality **Eleven v3**, the stable **Multilingual v2** supporting 29 languages, or the low-latency **Flash v2.5** for performance-critical applications. + +## Key capabilities + +- **Emotionally expressive speech** — Natural intonation, pacing, and tone across all three model tiers +- **Multi-language support** — 29 languages with Multilingual v2, broader coverage with Flash v2.5 +- **Low-latency option** — Flash v2.5 for real-time and performance-sensitive use cases +- **Voice customization** — Use ElevenLabs voice library or clone custom voices + + + + +## Available workflows + + + + Generate speech from text directly in your browser. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/elevenlabs/overview.mdx b/tutorials/partner-nodes/elevenlabs/overview.mdx new file mode 100644 index 000000000..9e7f2450b --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/overview.mdx @@ -0,0 +1,47 @@ +--- +title: "ElevenLabs Text-to-Speech, Speech-to-Text, Sound Effects & Voice Tools" +description: "Generate natural, emotionally rich speech, transcribe audio, create sound effects, and isolate voices using ElevenLabs in ComfyUI." +sidebarTitle: "ElevenLabs" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +ElevenLabs brings world-class voice AI to ComfyUI, covering text-to-speech, speech-to-text, sound effects, voice isolation, and voice transformation. Its **Eleven v3** and **Multilingual v2** models deliver emotionally expressive speech across 29 languages, while **Scribe v2** provides fast, accurate transcription with speaker diarization and entity detection. + +## Key capabilities + +- **Text to Speech** — Generate lifelike audio from text with three model tiers (Eleven v3, Multilingual v2, Flash v2.5) +- **Speech to Speech** — Transform a source audio clip into a different voice while keeping the original intonation and pacing +- **Speech to Text** — Transcribe audio with Scribe v2: 90+ languages, 32-speaker diarization, entity detection, keyword boost +- **Sound Effects** — Generate audio effects from text descriptions +- **Voice Isolation** — Separate voice from background noise, music, or ambient sound +- **Multi-Character Dialogue** — Generate conversations between multiple distinct voices in a single pass + + + + +## Models + +ElevenLabs offers a suite of specialized voice AI models, each optimized for a distinct task. From high-fidelity speech generation to real-time transcription and audio editing, the following models are available as workflow-powered nodes in ComfyUI. + + + + High-fidelity text-to-speech with three model tiers. + + + Fast, accurate speech-to-text with speaker diarization. + + + Transform audio into a different voice while preserving delivery. + + + Generate sound effects and ambient audio from text. + + + Multi-character dialogue with distinct voices in one pass. + + + Separate voice from background noise and music. + + diff --git a/tutorials/partner-nodes/elevenlabs/scribe-v2.mdx b/tutorials/partner-nodes/elevenlabs/scribe-v2.mdx new file mode 100644 index 000000000..879539cf3 --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/scribe-v2.mdx @@ -0,0 +1,31 @@ +--- +title: "ElevenLabs Speech-to-Text — Scribe v2" +description: "Transcribe audio to text with high accuracy, speaker diarization, and entity detection using ElevenLabs Scribe v2 in ComfyUI." +sidebarTitle: "Scribe v2" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Transcribe audio to text using **Scribe v2**, ElevenLabs' advanced speech-to-text model. It supports over 90 languages with speaker diarization for up to 32 speakers, entity detection, and keyword boosting for domain-specific terminology. + +## Key capabilities + +- **90+ language support** — Broad multilingual transcription coverage +- **Speaker diarization** — Up to 32 speakers identified and labeled +- **Entity detection** — Automatic identification of names, places, and key terms +- **Keyword boost** — Improve accuracy on domain-specific vocabulary + + + + +## Available workflows + + + + Transcribe audio to text directly in your workflow. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/elevenlabs/speech-to-speech.mdx b/tutorials/partner-nodes/elevenlabs/speech-to-speech.mdx new file mode 100644 index 000000000..fa42d4afc --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/speech-to-speech.mdx @@ -0,0 +1,30 @@ +--- +title: "ElevenLabs Speech to Speech" +description: "Transform an existing audio recording into a different voice or style while preserving content and delivery using ElevenLabs in ComfyUI." +sidebarTitle: "Speech to Speech" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Take an existing audio recording and transform it into a different voice or style while preserving the spoken content and delivery. Speech to Speech retains the original intonation, pacing, and emotional delivery while applying a new voice identity. + +## Key capabilities + +- **Preserve delivery** — Retain original intonation, pacing, and emotional expression +- **Voice transformation** — Apply a different voice identity to existing audio +- **Style transfer** — Change the vocal style or character without re-recording + + + + +## Available workflows + + + + Transform existing audio into a new voice. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/elevenlabs/text-to-dialogue.mdx b/tutorials/partner-nodes/elevenlabs/text-to-dialogue.mdx new file mode 100644 index 000000000..d5c721c30 --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/text-to-dialogue.mdx @@ -0,0 +1,30 @@ +--- +title: "ElevenLabs Text to Dialogue" +description: "Generate multi-character dialogue with distinct voices from a single text input using ElevenLabs in ComfyUI." +sidebarTitle: "Text to Dialogue" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Generate multi-character dialogue with distinct voices from a single text input. Define character voices and script the conversation — ElevenLabs renders each character with its designated voice in one pass. + +## Key capabilities + +- **Multi-voice generation** — Multiple characters with distinct voices in a single pass +- **Script-driven** — Input structured dialogue text with character assignments +- **Consistent character voices** — Each character maintains its designated voice identity + + + + +## Available workflows + + + + Generate multi-voice dialogue from text. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/elevenlabs/text-to-sound-effects.mdx b/tutorials/partner-nodes/elevenlabs/text-to-sound-effects.mdx new file mode 100644 index 000000000..68975f92f --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/text-to-sound-effects.mdx @@ -0,0 +1,30 @@ +--- +title: "ElevenLabs Text to Sound Effects" +description: "Generate sound effects, ambient audio, and foley from text descriptions using ElevenLabs in ComfyUI." +sidebarTitle: "Text to Sound Effects" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Generate sound effects, ambient audio, and foley from text descriptions. Describe the audio you need — from footsteps to thunderstorms — and ElevenLabs produces the matching sound effect. + +## Key capabilities + +- **Text-prompted sound generation** — Describe the sound, get the audio +- **Ambient environments** — Generate background atmospheres and room tones +- **Foley effects** — Create custom sound effects for video and game production + + + + +## Available workflows + + + + Generate sound effects from text. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/elevenlabs/voice-isolation.mdx b/tutorials/partner-nodes/elevenlabs/voice-isolation.mdx new file mode 100644 index 000000000..3343cead5 --- /dev/null +++ b/tutorials/partner-nodes/elevenlabs/voice-isolation.mdx @@ -0,0 +1,31 @@ +--- +title: "ElevenLabs Voice Isolation" +description: "Separate voice tracks from background noise, music, or ambient sound using ElevenLabs in ComfyUI." +sidebarTitle: "Voice Isolation" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Separate the voice track from background noise, music, or ambient sound in any audio recording. ElevenLabs Voice Isolation produces clean vocal tracks from noisy source material — recordings captured on-location, in crowds, or with background music. + +## Key capabilities + +- **Clean voice extraction** — Isolate speech from any background audio +- **Noise removal** — Eliminate ambient noise, wind, and room reverb +- **Music separation** — Separate vocals from music tracks +- **High preservation** — Maintain voice quality and naturalness during isolation + + + + +## Available workflows + + + + Isolate voice from noisy audio. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/grok/grok-imagine.mdx b/tutorials/partner-nodes/grok/grok-imagine.mdx new file mode 100644 index 000000000..c390a7319 --- /dev/null +++ b/tutorials/partner-nodes/grok/grok-imagine.mdx @@ -0,0 +1,51 @@ +--- +title: "Grok Imagine — Text to Image & Image Editing" +description: "Generate and edit images with xAI's Grok Imagine models — Pro, Standard, and Beta tiers, fast ~4s generation, cinematic quality with dramatic lighting and rich color saturation." +sidebarTitle: "Grok Imagine" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Generate images from text prompts and edit existing images using Grok Imagine by xAI, available in **Pro**, **Standard**, and **Beta** tiers. Grok delivers a distinct moody, cinematic aesthetic — dramatic lighting, rich color saturation, and strong character rendering with particular strengths in anime, portraiture, and cyberpunk styles. Image generation is notably fast at approximately **4 seconds per image**, making it well-suited for rapid iteration and exploration. + +Available output resolutions: **1K** and **2K**. Image editing supports 1 input image for Pro, or up to 3 input images for Standard/Beta. + +## Key capabilities + +- **Fast text-to-image generation** — ~4 seconds per image, ideal for rapid iteration +- **Image Editing** — Modify existing images with flexible aspect ratio options +- **Cinematic aesthetic** — Dramatic lighting, rich color saturation, moody atmosphere +- **Strong character rendering** — Excellent facial consistency, expressive lighting and framing +- **Anime & cyberpunk** — Stylized illustrations and neon compositions translate especially well + + + + +## Available workflows + +### Text to Image + +Generate images from text prompts using the highest quality model for pro-tier results. + + + + Generate images from text instantly. + + + Download the workflow JSON. + + + +### Image Editing + +Edit existing images using text-based instructions for modifications. + + + + Edit images with text instructions. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/grok/grok-video.mdx b/tutorials/partner-nodes/grok/grok-video.mdx new file mode 100644 index 000000000..d03e355d0 --- /dev/null +++ b/tutorials/partner-nodes/grok/grok-video.mdx @@ -0,0 +1,78 @@ +--- +title: "Grok Video — Text to Video, Video Editing, Extending & Reference to Video" +description: "Generate, edit, extend, and create video from reference images using xAI's Grok Video models in ComfyUI." +sidebarTitle: "Grok Video" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Generate video from text descriptions, edit existing video footage, extend clips naturally, and create consistent video using reference images — all with Grok's signature moody visual style. Grok Video brings xAI's cinematic aesthetic to motion content. The model is particularly effective at motion graphics, 2D illustrations, and image-to-video world rebuilding from new angles. + +Available in Standard and Beta tiers. + +## Key capabilities + +- **Text to Video** — Generate video from text descriptions with cinematic quality +- **Video Editing** — Edit existing video footage with text instructions (input: 1-8.7s, max 50MB) +- **Video Extending** — Extend an input video of up to 15 seconds by up to 10 seconds with natural continuations (~60s generation time for 10s extension) +- **Reference to Video** — Generate video guided by up to 7 reference images for identity consistency (~100s for 10s video at 720p) +- **Motion graphics & illustrations** — Flat and graphic inputs animate cleanly into bold video results +- **Image-to-video world rebuilding** — Reconstructs scenes from new angles or perspectives + + + + +## Available workflows + +### Text to Video + +Generate video from text descriptions with Grok's signature visual style. + + + + Generate video from text. + + + Download the workflow JSON. + + + +### Video Editing + +Edit existing video footage using text instructions. + + + + Edit existing video with text. + + + Download the workflow JSON. + + + +### Video Extending + +Extend an existing video clip by up to 10 seconds, creating natural continuations rather than loops. + + + + Extend a video clip. + + + Download the workflow JSON. + + + +### Reference to Video + +Generate consistent video content guided by up to 7 reference images for identity, scene, and narrative consistency. + + + + Generate video with reference images. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/grok/overview.mdx b/tutorials/partner-nodes/grok/overview.mdx new file mode 100644 index 000000000..d6237f2e6 --- /dev/null +++ b/tutorials/partner-nodes/grok/overview.mdx @@ -0,0 +1,37 @@ +--- +title: "Grok Image & Video Generation" +description: "Generate images and videos with xAI's Grok models — fast ~4s image generation, moody cinematic aesthetics, with strong anime and character rendering. Available in Pro, Standard, and Beta tiers." +sidebarTitle: "Grok" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Grok Imagine by xAI brings a distinct, moody visual style to AI image and video generation. It produces cinematic-quality images with dramatic lighting and rich color saturation, and supports text-to-video, video editing, and reference-to-video generation with consistent character identity. Image generation is notably fast at approximately 4 seconds per image, ideal for rapid iteration. + +Grok performs particularly well with anime, cyberpunk, and cinematic portrait styles, and excels at reconstructing scenes from new angles in image-to-video workflows. + +## Key capabilities + +- **Text to Image** — Generate images from text prompts (~4 seconds per image) +- **Image Editing** — Edit existing images with text instructions +- **Text to Video** — Generate video from text descriptions +- **Video Editing** — Edit existing video footage +- **Video Extending** — Extend an input video (15s in, extend up to 10s, ~60s generation time) +- **Reference to Video** — Generate consistent video using up to 7 reference images (~100s for 10s at 720p) + + + + +## Models + +Grok's creative tools are organized into two primary model groups. **Grok Imagine** handles all image generation and editing tasks, while **Grok Video** covers video generation, editing, extension, and reference-based video creation. + + + + Image generation and editing with text prompts. + + + Video generation, editing, and extension with reference support. + + diff --git a/tutorials/partner-nodes/hitpaw/hitpaw-image-enhance.mdx b/tutorials/partner-nodes/hitpaw/hitpaw-image-enhance.mdx new file mode 100644 index 000000000..d16f61299 --- /dev/null +++ b/tutorials/partner-nodes/hitpaw/hitpaw-image-enhance.mdx @@ -0,0 +1,27 @@ +--- +title: "HitPaw Image Enhance" +description: "Upscale and restore images using HitPaw's general AI enhancement model in ComfyUI." +sidebarTitle: "HitPaw Image Enhance" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +HitPaw Image Enhance is a multi-purpose AI upscaling and restoration model that handles noise, blur, artifacts, and low resolution across photographs, renders, and AI-generated images. It uses content-aware processing to automatically apply optimized enhancement strategies for natural results. + +## Key capabilities + +- **Multi-purpose upscaling** — Handles noise, blur, artifacts, and low resolution in a single pass. +- **Content-aware processing** — Automatically detects image content type (portrait, landscape, text, etc.) and applies optimized strategies. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure. +- **Minimal configuration** — Simple drop-in node with reliable results and no extensive parameter tuning. + + + + +## Available workflows + + + Launch the image enhancement workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/hitpaw/hitpaw-video-enhance.mdx b/tutorials/partner-nodes/hitpaw/hitpaw-video-enhance.mdx new file mode 100644 index 000000000..2b0bf05fe --- /dev/null +++ b/tutorials/partner-nodes/hitpaw/hitpaw-video-enhance.mdx @@ -0,0 +1,27 @@ +--- +title: "HitPaw Video Enhance" +description: "Enhance and upscale video footage using HitPaw's AI video enhancement model in ComfyUI." +sidebarTitle: "HitPaw Video Enhance" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +HitPaw Video Enhance applies frame-by-frame AI processing to upscale and denoise video footage. It improves resolution, reduces noise, and maintains temporal smoothness across frames — suitable for compressed footage, screen recordings, and archival video. + +## Key capabilities + +- **Frame-by-frame upscaling** — AI-powered video upscaling with temporal consistency. +- **Denoising & restoration** — Removes noise and artifacts from compressed or archival footage. +- **Temporal smoothness** — Maintains consistent quality across frames without flickering or artifacts. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure. + + + + +## Available workflows + + + Launch the video enhancement workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/hitpaw/overview.mdx b/tutorials/partner-nodes/hitpaw/overview.mdx new file mode 100644 index 000000000..789db588d --- /dev/null +++ b/tutorials/partner-nodes/hitpaw/overview.mdx @@ -0,0 +1,35 @@ +--- +title: "Image & Video Enhancement with HitPaw" +description: "Enhance and upscale images and video using HitPaw's AI models within ComfyUI — covering general image enhancement and video enhancement workflows." +sidebarTitle: "HitPaw" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +HitPaw brings production-grade AI image and video enhancement to ComfyUI as native nodes. Its **image enhancement** and **video enhancement** models handle upscaling, denoising, and restoration with minimal configuration, all powered by ComfyCloud with no local GPU needed. + +## Key capabilities + +- **General image enhancement** — Multi-purpose AI upscaling and restoration that handles noise, blur, artifacts, and low resolution across photographs, renders, and AI-generated images. +- **Video enhancement** — Frame-by-frame video upscaling and denoising with temporal consistency, suitable for compressed footage, screen recordings, and archival video. +- **Content-aware processing** — Automatically detects image content type (portrait, landscape, text, etc.) and applies optimized enhancement strategies for natural results. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure, freeing your local hardware for other tasks. +- **Minimal configuration** — Simple drop-in nodes that produce reliable results without extensive parameter tuning, ideal for integration into larger ComfyUI pipelines. +- **Broad format support** — Compatible with common image and video formats, making it a versatile post-processing step for any ComfyUI workflow. + + + + +## Models + +HitPaw offers two AI enhancement workflows in ComfyUI — one for images and one for video. Both are designed as simple drop-in nodes with minimal configuration, suitable for integration into any pipeline. + + + + Multi-purpose AI image upscaling and restoration. + + + Frame-by-frame video upscaling and denoising. + + diff --git a/tutorials/partner-nodes/magnific/magnific-creative-tools.mdx b/tutorials/partner-nodes/magnific/magnific-creative-tools.mdx new file mode 100644 index 000000000..df495ce46 --- /dev/null +++ b/tutorials/partner-nodes/magnific/magnific-creative-tools.mdx @@ -0,0 +1,60 @@ +--- +title: "Magnific Creative Tools | ComfyUI Partner Nodes" +description: "Learn how to use Magnific Creative Tools in ComfyUI for Style Transfer, Relighting, and Skin Enhancement." +sidebarTitle: "Magnific Creative Tools" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Magnific Creative Tools bring studio-quality post-processing to ComfyUI, including **Style Transfer** using reference images, **Relighting** with custom direction and color temperature, and **Skin Enhancement** for intelligent portrait retouching. + + + + +## Key capabilities + +- **Style Transfer** — Transfers the artistic style, color palette, and lighting from a reference image onto your source image for creative transformations. +- **Relight** — Adjusts lighting direction, intensity, and color temperature on existing images, enabling seamless compositing into new environments. +- **Skin Enhancer** — Intelligent portrait retouching that smooths skin, reduces blemishes, and evens skin tone while preserving natural facial features. + +## Available workflows + +### Style Transfer + +Apply the visual style, color palette, and lighting mood from a reference image to your source image. This enables artistic reinterpretations — turn a photograph into an oil painting, or match the aesthetic of a cinematic frame. + + + + Run the style transfer workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Relight + +Adjust the lighting on your image by specifying direction, intensity, and color temperature. This is especially useful for compositing subjects into new scenes where the original lighting doesn't match. + + + + Run the relight workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Skin Enhancer + +Apply intelligent portrait retouching to smooth skin, reduce blemishes, and even out skin tone while preserving natural facial features and texture. + + + + Run the skin enhancer workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + diff --git a/tutorials/partner-nodes/magnific/magnific-upscale.mdx b/tutorials/partner-nodes/magnific/magnific-upscale.mdx new file mode 100644 index 000000000..d66bff005 --- /dev/null +++ b/tutorials/partner-nodes/magnific/magnific-upscale.mdx @@ -0,0 +1,46 @@ +--- +title: "Magnific Upscale | ComfyUI Partner Nodes" +description: "Learn how to use Magnific Upscale in ComfyUI for Precise and Creative image upscaling up to 16x magnification." +sidebarTitle: "Magnific Upscale" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Magnific Upscale offers two modes for enlarging your images: **Precise** for high-fidelity detail preservation, and **Creative** for AI-generated detail at up to 16x magnification. + + + + +## Key capabilities + +- **Precise Upscale** — High-fidelity image enlargement that preserves original details and introduces minimal new content; ideal for archival or product images. +- **Creative Upscale (16x)** — AI-powered upscaling that interprets and adds realistic detail at up to 16x magnification; the Creativity slider controls how much new detail is generated. + +## Available workflows + +### Image Upscale — Precise + +Generate a high-fidelity upscaled version of your input image with precise detail preservation. This mode adds minimal new content and stays as close to the original as possible. + + + + Run the precise image upscale workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Image Upscale — Creative + +Generate a creatively upscaled version of your input image at up to 16x magnification. The Creativity slider allows you to control how much new AI-generated detail is added — lower values stay closer to the original, while higher values produce richer, more detailed results. + + + + Run the creative image upscale workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + diff --git a/tutorials/partner-nodes/magnific/overview.mdx b/tutorials/partner-nodes/magnific/overview.mdx new file mode 100644 index 000000000..713971b4f --- /dev/null +++ b/tutorials/partner-nodes/magnific/overview.mdx @@ -0,0 +1,32 @@ +--- +title: "Magnific AI: Image Enhancement & Creative Tools | ComfyUI Partner Nodes" +description: "This guide covers how to use the Magnific partner nodes in ComfyUI for AI-powered image enhancement, upscaling, style transfer, relighting, and skin enhancement." +sidebarTitle: "Magnific AI" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Magnific AI is a professional-grade image enhancement platform, now natively integrated into ComfyUI. It offers **Precise** and **Creative** upscaling (up to 16x), style transfer, relighting, and skin enhancement — bringing studio-quality post-processing directly into your workflow. + + + + +## Key capabilities + +- **Precise Upscale** — High-fidelity image enlargement that preserves original details and introduces minimal new content; ideal for archival or product images. +- **Creative Upscale (16x)** — AI-powered upscaling that interprets and adds realistic detail at up to 16x magnification; the Creativity slider controls how much new detail is generated. +- **Style Transfer** — Transfers the artistic style, color palette, and lighting from a reference image onto your source image for creative transformations. +- **Relight** — Adjusts lighting direction, intensity, and color temperature on existing images, enabling seamless compositing into new environments. +- **Skin Enhancer** — Intelligent portrait retouching that smooths skin, reduces blemishes, and evens skin tone while preserving natural facial features. + +## Models + + + + Upscale your images with Precise (detail-preserving) or Creative (AI-generated detail at up to 16x). + + + Apply Style Transfer, Relighting, and Skin Enhancement for studio-quality creative edits. + + diff --git a/tutorials/partner-nodes/minimax/minimax-01.mdx b/tutorials/partner-nodes/minimax/minimax-01.mdx new file mode 100644 index 000000000..c11f3c5d1 --- /dev/null +++ b/tutorials/partner-nodes/minimax/minimax-01.mdx @@ -0,0 +1,48 @@ +--- +title: "MiniMax-01 - Text and image to video" +description: "Generate cinematic video from text or images using MiniMax-01 with Director mode and Live motion in ComfyUI" +sidebarTitle: "MiniMax-01" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +MiniMax-01 is the first-generation MiniMax architecture, encompassing T2V-01 (text-to-video) and I2V-01 (image-to-video) models. T2V-01 offers full Director mode with 15 cinematic camera directions, while I2V-01 provides two modes — Director for camera-controlled image animation and Live for natural, organic motion that brings portraits, landscapes, and product shots to life. + +## Key capabilities + +- **T2V-01 Director mode** — 15 camera movement types including pan, tilt, dolly, crane, tracking, aerial, handheld, roll, zoom-in, zoom-out, push-in, pull-out, orbit, first-person, and static +- **I2V-01-Director** — Apply cinematic camera controls to image-initiated videos +- **I2V-01-Live** — Natural, lively motion with organic movement dynamics from a single image +- **High motion quality** — Industry-leading realism in character movement, physics, and fluid dynamics + + + + +## Available workflows + +### Text to video (T2V-01) + +Generate a video from a text prompt with full Director mode support. Choose from 15 camera movement types to craft the exact cinematic feel you need. + + + + Try the T2V-01 workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Image to video (I2V-01) + +Animate a starting image into a video clip. Two modes are available: I2V-01-Director brings camera direction (pan, zoom, tilt, etc.) to image-based generation, while I2V-01-Live prioritizes natural, organic motion. + + + + Try the I2V-01 workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/minimax/minimax-s2v-01.mdx b/tutorials/partner-nodes/minimax/minimax-s2v-01.mdx new file mode 100644 index 000000000..ab1411295 --- /dev/null +++ b/tutorials/partner-nodes/minimax/minimax-s2v-01.mdx @@ -0,0 +1,31 @@ +--- +title: "MiniMax S2V-01 - Subject reference to video" +description: "Generate video while maintaining character, face, or object identity using MiniMax S2V-01 subject reference in ComfyUI" +sidebarTitle: "MiniMax S2V-01" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +MiniMax S2V-01 enables subject reference-to-video generation, maintaining the identity of a specific character, face, or object throughout the clip. Supply a reference image alongside your text prompt, and S2V-01 ensures consistent appearance across the entire generated sequence — ideal for character-driven narratives, branded content, and product showcases where visual identity must remain stable from frame to frame. + +## Key capabilities + +- **Subject identity preservation** — Maintain character, face, or object identity across the entire generated video +- **Single reference image** — One input image is sufficient to guide consistent appearance +- **Text-guided generation** — Combine subject reference with text prompts for full creative control +- **Perfect for serial content** — Ideal for character-driven narratives, branded content, and product showcases + + + + +## Available workflows + + + + Try the S2V-01 workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/minimax/overview.mdx b/tutorials/partner-nodes/minimax/overview.mdx new file mode 100644 index 000000000..50b33cd22 --- /dev/null +++ b/tutorials/partner-nodes/minimax/overview.mdx @@ -0,0 +1,32 @@ +--- +title: "MiniMax (Hailuo) - AI video generation" +description: "Generate high-quality video from text, images, and subject references using MiniMax's T2V-01, I2V-01, and S2V-01 models in ComfyUI" +sidebarTitle: "MiniMax (Hailuo)" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +MiniMax (Hailuo) is a leading AI video generation platform renowned for exceptional motion quality and cinematic output, now inside ComfyUI. It offers two model lines — **MiniMax-01** (T2V-01 text-to-video and I2V-01 image-to-video with Director/Live modes) and **MiniMax S2V-01** (subject reference-to-video) — covering everything from film-grade narrative to character-driven content. + +## Key capabilities + +- **T2V-01 with Director mode** — 15 cinematic camera movement types (pan, tilt, dolly, crane, tracking, aerial, handheld, roll, zoom-in, zoom-out, push-in, pull-out, orbit, first-person, static) for precise shot-level direction +- **I2V-01-Director** — Apply cinematic camera controls to image-initiated videos, combining the visual foundation of a starting image with the expressive language of Director mode +- **I2V-01-Live** — Generate natural, lively motion from a single image with organic movement dynamics — ideal for bringing portraits, landscapes, and product shots to life +- **S2V-01 subject reference** — Maintain character, face, or object identity across the generated video using a single reference image; perfect for serial content +- **High motion quality** — Industry-leading realism in character movement, physics, and fluid dynamics across all model variants + + + + +## Models + + + + T2V-01 with Director mode and I2V-01 with Director/Live modes — the first-generation MiniMax architecture for text and image-based video generation. + + + Subject reference-to-video for maintaining character, face, or object identity across generated clips. + + diff --git a/tutorials/partner-nodes/pixverse/overview.mdx b/tutorials/partner-nodes/pixverse/overview.mdx new file mode 100644 index 000000000..42cbef9cc --- /dev/null +++ b/tutorials/partner-nodes/pixverse/overview.mdx @@ -0,0 +1,27 @@ +--- +title: "PixVerse AI: Video Generation | ComfyUI Partner Nodes" +description: "This guide covers how to use the PixVerse partner nodes in ComfyUI for AI-powered video generation from text prompts, images, and template-based workflows." +sidebarTitle: "PixVerse AI" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +PixVerse is an AI video generation platform that produces cinematic videos from text and images, now integrated as partner nodes in ComfyUI. With support for **V5.6**, **V6**, **C1**, and **R1** models, it enables **Text-to-Video**, **Image-to-Video**, and **Template Image-to-Video** modes for rapid, repeatable video creation without separate API keys or external setup. + + + + +## Key capabilities + +- **Text-to-Video** — Generate original videos from text prompts, ideal for concept visualization, storyboarding, and rapid creative exploration. +- **Image-to-Video** — Animate a reference image while preserving subject identity and scene composition, bringing static visuals to life with natural motion. +- **Template Image-to-Video** — Apply pre-defined motion templates to input images for predictable, repeatable animation styles suitable for production pipelines. + +## Models + + + + Generate videos from text prompts, images, or templates using PixVerse's AI models. + + diff --git a/tutorials/partner-nodes/pixverse/pixverse-video.mdx b/tutorials/partner-nodes/pixverse/pixverse-video.mdx new file mode 100644 index 000000000..9f4e93d89 --- /dev/null +++ b/tutorials/partner-nodes/pixverse/pixverse-video.mdx @@ -0,0 +1,60 @@ +--- +title: "PixVerse Video | ComfyUI Partner Nodes" +description: "Learn how to use PixVerse Video in ComfyUI for Text-to-Video, Image-to-Video, and Template Image-to-Video generation." +sidebarTitle: "PixVerse Video" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +PixVerse Video enables cinematic video creation directly in ComfyUI through **Text-to-Video**, **Image-to-Video**, and **Template Image-to-Video** modes, powered by models including V5.6, V6, C1, and R1. + + + + +## Key capabilities + +- **Text-to-Video** — Generate original videos from text prompts, ideal for concept visualization, storyboarding, and rapid creative exploration. +- **Image-to-Video** — Animate a reference image while preserving subject identity and scene composition, bringing static visuals to life with natural motion. +- **Template Image-to-Video** — Apply pre-defined motion templates to input images for predictable, repeatable animation styles suitable for production pipelines. + +## Available workflows + +### Text-to-Video + +Create original videos from text prompts. Describe the scene, characters, and action — PixVerse generates a corresponding video clip with cinematic quality. + + + + Run the text-to-video workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Image-to-Video + +Take a reference image and generate a video that animates it while preserving the original subject, style, and composition. + + + + Run the image-to-video workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Template Image-to-Video + +Apply a pre-defined motion template to your input image for predictable, repeatable animation results. This mode reduces variability and is ideal for consistent batch processing. + + + + Run the template image-to-video workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + diff --git a/tutorials/partner-nodes/quiver/overview.mdx b/tutorials/partner-nodes/quiver/overview.mdx new file mode 100644 index 000000000..ac7bf08fc --- /dev/null +++ b/tutorials/partner-nodes/quiver/overview.mdx @@ -0,0 +1,29 @@ +--- +title: "Quiver AI: SVG Generation | ComfyUI Partner Nodes" +description: "This guide covers how to use the Quiver partner nodes in ComfyUI for AI-powered SVG vector graphics generation from text prompts and reference images." +sidebarTitle: "Quiver AI" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Quiver is an AI-powered vector graphics platform that generates high-quality SVGs from text descriptions and raster images, now available as partner nodes in ComfyUI. With models from **Arrow 1.1** (fast) to **Arrow 1.1 Max** (high-fidelity) and **Arrow Preview** (experimental), it produces resolution-independent vector artwork without manual editing. + + + + +## Key capabilities + +- **Text-to-SVG** — Generate fully editable SVG vector graphics from natural language descriptions; perfect for icons, logos, illustrations, and UI elements. +- **Image-to-SVG** — Convert raster images (PNG, JPG) into clean, scalable SVG vector graphics by tracing shapes, colors, and gradients. +- **Arrow 1.1** — Standard model offering fast and efficient SVG generation suitable for most use cases. +- **Arrow 1.1 Max** — High-quality variant for detailed and accurate vector output on complex illustrations. +- **Arrow Preview** — Early access to experimental features and upcoming model improvements. + +## Models + + + + Generate editable SVG vector graphics from text descriptions or convert raster images into scalable vector art. + + diff --git a/tutorials/partner-nodes/quiver/quiver-svg.mdx b/tutorials/partner-nodes/quiver/quiver-svg.mdx new file mode 100644 index 000000000..e0cac18c2 --- /dev/null +++ b/tutorials/partner-nodes/quiver/quiver-svg.mdx @@ -0,0 +1,46 @@ +--- +title: "Quiver SVG | ComfyUI Partner Nodes" +description: "Learn how to use Quiver SVG in ComfyUI for Text-to-SVG and Image-to-SVG vector graphics generation." +sidebarTitle: "Quiver SVG" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Quiver SVG generates resolution-independent vector graphics directly in ComfyUI through **Text-to-SVG** and **Image-to-SVG** modes. With models including Arrow 1.1, Arrow 1.1 Max, and Arrow Preview, it handles everything from quick icon generation to complex illustration vectorization. + + + + +## Key capabilities + +- **Text-to-SVG** — Generate fully editable SVG vector graphics from natural language descriptions; perfect for icons, logos, illustrations, and UI elements. +- **Image-to-SVG** — Convert raster images (PNG, JPG) into clean, scalable SVG vector graphics by tracing shapes, colors, and gradients. + +## Available workflows + +### Text-to-SVG + +Describe the vector graphic you want in natural language, and Quiver generates a clean, editable SVG. Use it for creating icons, logos, diagrams, illustrations, and UI components. + + + + Run the text-to-SVG workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + + +### Image-to-SVG + +Convert an existing raster image into a clean, scalable SVG vector graphic. Shapes, colors, and gradients from the source are traced and reconstructed as vector paths. + + + + Run the image-to-SVG workflow instantly on Comfy Cloud. + + + Download the workflow JSON file for local use. + + diff --git a/tutorials/partner-nodes/topaz/overview.mdx b/tutorials/partner-nodes/topaz/overview.mdx new file mode 100644 index 000000000..765d3c8f5 --- /dev/null +++ b/tutorials/partner-nodes/topaz/overview.mdx @@ -0,0 +1,36 @@ +--- +title: "Image & Video Enhancement with Topaz Labs" +description: "Upscale, enhance, and restore images and video using Topaz Labs AI models — including Reimagine creative upscaling, face enhancement, Starlight diffusion upscale, and Apollo frame interpolation — all within ComfyUI." +sidebarTitle: "Topaz Labs" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Topaz Labs brings enterprise-grade AI image and video enhancement to ComfyUI — trusted by professionals worldwide. Its suite covers **image enhancement**, **landscape upscaling**, **illustration upscaling**, and **video enhancement**, all powered by ComfyCloud with no local GPU needed. + +## Key capabilities + +- **Reimagine creative upscaling** — Guide the upscaling process with text prompts to creatively reinterpret and enhance images, not just enlarge them. +- **Face enhancement** — Dedicated facial detail reconstruction model that detects and sharpens faces in images and video, preserving natural skin texture and expression. +- **8K output support** — Upscale images and video frames to ultra-high 8K resolution while maintaining sharpness and visual coherence. +- **Starlight diffusion upscaler** — Diffusion-based upscaling optimized for landscape photography, hallucinating realistic detail like foliage, rock textures, and clouds. +- **Apollo frame interpolation** — Generate smooth intermediate frames between existing video frames for slow-motion effects or frame rate conversion. +- **Illustration-optimized upscale** — Specialized model for flat-colored artwork, line art, and digital illustrations that preserves hard edges and clean gradients. +- **Video enhancement pipeline** — Comprehensive video processing including denoising, deinterlacing, stabilization, and intelligent frame-by-frame upscaling. + + + + +## Models + +Topaz Labs offers a suite of AI enhancement models for both image and video. The **Image Enhance** group covers general upscaling, landscape diffusion upscaling, and illustration-optimized upscaling, while **Video Enhance** provides frame-by-frame enhancement with Apollo interpolation. + + + + General image upscaling, Starlight landscape diffusion, and illustration upscaling. + + + Frame-by-frame video upscaling, denoising, and Apollo interpolation. + + diff --git a/tutorials/partner-nodes/topaz/topaz-image.mdx b/tutorials/partner-nodes/topaz/topaz-image.mdx new file mode 100644 index 000000000..c90f1d77d --- /dev/null +++ b/tutorials/partner-nodes/topaz/topaz-image.mdx @@ -0,0 +1,50 @@ +--- +title: "Topaz Image Enhance" +description: "Upscale, enhance, and restore images using Topaz Labs AI — including general image enhancement, Starlight landscape diffusion upscaling, and illustration-optimized upscaling in ComfyUI." +sidebarTitle: "Topaz Image Enhance" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Topaz Image Enhance offers three specialized workflows for AI-powered image upscaling and restoration. The **general enhancement** model handles photographs and renders, the **Starlight landscape upscaler** uses diffusion for photorealistic landscape detail, and the **illustration upscale** preserves hard edges and clean gradients for digital art. + +## Key capabilities + +- **General-purpose enhancement** — AI upscaling and denoising suitable for photographs, renders, and most image types. +- **Starlight landscape diffusion** — Diffusion-based upscaling that hallucinates realistic natural detail like foliage, rock textures, and clouds at up to 8K resolution. +- **Illustration-optimized upscaling** — Specialized model for flat-colored artwork, line art, and digital illustrations that preserves hard edges and clean color gradients. +- **Face enhancement** — Dedicated facial detail reconstruction that preserves natural skin texture and expression. +- **8K output support** — Upscale images to ultra-high 8K resolution while maintaining sharpness and visual coherence. + + + + +## Available workflows + +### Image Enhancement + +General-purpose AI image upscaling and denoising using Topaz's core enhancement model. Suitable for photographs, renders, and most image types. + + + Launch the image enhancement workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + + +### Landscape Upscaler (Starlight) + +Diffusion-based upscaling specialized for landscape photography. Uses the Starlight model to hallucinate realistic natural detail at up to 8K resolution. + + + Launch the Starlight landscape upscaler on ComfyCloud + Download the workflow JSON for local ComfyUI + + +### Illustration Upscale + +Optimized upscaling for digital illustrations, anime art, and flat-color artwork. Preserves hard edges, crisp lines, and smooth color gradients. + + + Launch the illustration upscale workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/topaz/topaz-video-enhance.mdx b/tutorials/partner-nodes/topaz/topaz-video-enhance.mdx new file mode 100644 index 000000000..140a3bfba --- /dev/null +++ b/tutorials/partner-nodes/topaz/topaz-video-enhance.mdx @@ -0,0 +1,28 @@ +--- +title: "Topaz Video Enhance" +description: "Enhance, upscale, and interpolate video using Topaz Labs AI — including frame-by-frame upscaling, denoising, deinterlacing, stabilization, and Apollo frame interpolation in ComfyUI." +sidebarTitle: "Topaz Video Enhance" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Topaz Video Enhance applies AI-powered frame-by-frame processing to upscale, denoise, and restore video footage. It supports deinterlacing, stabilization, and the Apollo frame interpolation model for creating smooth slow-motion effects. + +## Key capabilities + +- **Frame-by-frame upscaling** — Intelligent AI upscaling that maintains consistency across video frames. +- **Denoising & restoration** — Removes noise, artifacts, and compression artifacts from video footage. +- **Deinterlacing & stabilization** — Converts interlaced video and stabilizes shaky footage. +- **Apollo frame interpolation** — Generates smooth intermediate frames for slow-motion effects or frame rate conversion. +- **8K output support** — Upscale video frames to ultra-high 8K resolution. + + + + +## Available workflows + + + Launch the video enhancement workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/veo/overview.mdx b/tutorials/partner-nodes/veo/overview.mdx new file mode 100644 index 000000000..280b6a58f --- /dev/null +++ b/tutorials/partner-nodes/veo/overview.mdx @@ -0,0 +1,35 @@ +--- +title: "Video Generation with Google Veo" +description: "Generate cinematic, high-resolution videos using Google DeepMind's Veo models — Veo 3.1 and Veo 2.0 — directly within ComfyUI." +sidebarTitle: "Veo" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Google DeepMind's **Veo 3.1** and **Veo 2.0** bring state-of-the-art AI video generation to ComfyUI. Veo 3.1 delivers cinema-grade 4K video with native audio, while Veo 2.0 offers fast image-to-video for rapid prototyping — all seamlessly integrated into your ComfyUI pipeline. + +## Key capabilities + +- **Cinema-grade 4K video** — Veo 3.1 generates 4K resolution video with crisp detail, natural lighting, and realistic physics, suitable for professional film and advertising use. +- **Native audio sync** — Video output includes synchronized audio, eliminating the need for separate audio generation and manual alignment in post-production. +- **Text-to-video (Veo 3.1)** — Generate high-quality video directly from text prompts, with support for detailed scene direction, camera movement, and visual style control. +- **Optimized speed tier** — A faster inference variant of Veo 3.1 that trades slight quality for significantly reduced generation time, ideal for iteration and draft production. +- **Image-to-video (Veo 2.0)** — Animate static images into short video clips, preserving the original composition while adding natural motion. +- **Seamless ComfyUI pipeline** — Full integration allows chaining Veo output with ComfyUI's existing upscaling, audio, and compositing nodes without file export steps. + + + + +## Models + +Google Veo offers two model generations in ComfyUI, each optimized for different use cases. Veo 3.1 is the flagship text-to-video model with two speed tiers, while Veo 2.0 provides a lightweight image-to-video option. + + + + Cinema-grade 4K text-to-video with native audio. Standard and Fast tiers available. + + + Lightweight image-to-video for animating static images. + + diff --git a/tutorials/partner-nodes/veo/veo-2-0.mdx b/tutorials/partner-nodes/veo/veo-2-0.mdx new file mode 100644 index 000000000..0f2be174c --- /dev/null +++ b/tutorials/partner-nodes/veo/veo-2-0.mdx @@ -0,0 +1,26 @@ +--- +title: "Veo 2.0 — Image-to-Video" +description: "Animate static images into short video clips using Google DeepMind's Veo 2.0 in ComfyUI." +sidebarTitle: "Veo 2.0" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Veo 2.0 is Google DeepMind's lightweight image-to-video model in ComfyUI. It animates static images into short video clips, preserving the original composition while adding natural motion — ideal for breathing life into concept art, AI-generated images, or photographs. + +## Key capabilities + +- **Image-to-video animation** — Convert static images into short, natural-looking video clips. +- **Composition preservation** — Maintains the original image structure, colors, and layout while adding motion. +- **Natural motion generation** — Produces realistic movement that respects the source image's content and style. + + + + +## Available workflows + + + Launch the Veo 2.0 I2V workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/veo/veo-3-1.mdx b/tutorials/partner-nodes/veo/veo-3-1.mdx new file mode 100644 index 000000000..e9517d37f --- /dev/null +++ b/tutorials/partner-nodes/veo/veo-3-1.mdx @@ -0,0 +1,40 @@ +--- +title: "Veo 3.1 — Text-to-Video" +description: "Generate cinema-grade 4K video with native audio using Google DeepMind's Veo 3.1 in ComfyUI — available in Standard and Fast tiers." +sidebarTitle: "Veo 3.1" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Veo 3.1 is Google DeepMind's flagship text-to-video model, delivering cinema-grade 4K video with synchronized native audio. Available in **Standard** (full quality) and **Fast** (optimized speed) tiers, it supports detailed scene direction, camera movement, and visual style control. + +## Key capabilities + +- **Cinema-grade 4K video** — Generates 4K resolution video with crisp detail, natural lighting, and realistic physics. +- **Native audio sync** — Video output includes synchronized audio, eliminating the need for post-production alignment. +- **Text-to-video generation** — Create high-quality video directly from text prompts with full creative control. +- **Optimized Fast tier** — A faster inference variant that trades slight quality for significantly reduced generation time, ideal for iteration and draft production. + + + + +## Available workflows + +### Veo 3.1 — Text-to-Video (Standard) + +Generate full-quality 4K video with native audio from a text prompt. Best for final production output where quality matters most. + + + Launch the full-quality Veo 3.1 workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + + +### Veo 3.1 — Text-to-Video (Fast) + +A lightweight variant of Veo 3.1 optimized for speed. Use this for rapid ideation, draft generation, and iterative prompt refinement. + + + Launch the fast Veo 3.1 workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/vidu/overview.mdx b/tutorials/partner-nodes/vidu/overview.mdx new file mode 100644 index 000000000..b296e788d --- /dev/null +++ b/tutorials/partner-nodes/vidu/overview.mdx @@ -0,0 +1,36 @@ +--- +title: "Vidu - AI video generation" +description: "Generate high-quality video from text, images, and reference materials using Vidu's Q1, Q2, and Q3 model generations in ComfyUI" +sidebarTitle: "Vidu" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Vidu is a fast AI video generation platform by Shengshu Technology, now available inside ComfyUI. Its three model generations — **Q1**, **Q2**, and **Q3** — deliver a range of speed and quality options, from foundation generation with broad input support to 3x faster inference with camera control and exceptional anime output driven by the latest Vidu engines (2.0, 3.0, and 3.5). + +## Key capabilities + +- **Multi-generation evolution** — Choose between Q1 (foundation), Q2 (fast + multi-reference + camera control), and Q3 (latest narrative + anime/2D focus) to match your workflow +- **Fast inference** — Q3 generates 4-second clips in as fast as 10 seconds; Q2 delivers 3x the speed of Q1 +- **Rich input modes** — Text-to-video, image-to-video, reference-to-video (up to 7 images), start-end-to-video, video extension, and first-last-frame-to-video +- **Camera language (Q2)** — Pan, zoom, rotation, and other cinematic moves for precise shot direction +- **Exceptional anime & 2D animation (Q3)** — Industry-leading output quality for animation production pipelines +- **Subject consistency** — Maintain character and scene identity across multi-reference inputs + + + + +## Models + + + + Foundation generation with text-to-video, image-to-video, reference-to-video and more — the versatile starting point. + + + 3x faster inference with multi-reference support and cinematic camera control — ideal for iterative workflows. + + + The latest generation with improved narrative understanding, outstanding anime/2D quality, and Vidu 3.5 engine. + + diff --git a/tutorials/partner-nodes/vidu/vidu-q1.mdx b/tutorials/partner-nodes/vidu/vidu-q1.mdx new file mode 100644 index 000000000..45339f053 --- /dev/null +++ b/tutorials/partner-nodes/vidu/vidu-q1.mdx @@ -0,0 +1,77 @@ +--- +title: "Vidu Q1 - Foundation video generation" +description: "Generate video from text, images, reference images, start-end frames, and extend existing clips using Vidu Q1 in ComfyUI" +sidebarTitle: "Vidu Q1" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Vidu Q1 is the first-generation Vidu model that handles all core video generation types. It offers a broad range of input modes — text, image, reference images, start-and-end frames, and video extension — making it a reliable choice for general-purpose production where speed and expressiveness are balanced. + +## Key capabilities + +- **Five generation modes** — Text-to-video, image-to-video, reference-to-video (up to 7 images), start-end-to-video, and video extension +- **Flexible subject reference** — Maintain character and scene identity across multi-reference inputs +- **Video extension** — Extend existing video clips with coherent continuation +- **Balanced performance** — Reliable quality suitable for a wide range of production scenarios + + + + +## Available workflows + +### Text to video + + + + Try the Q1 Text-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Image to video + + + + Try the Q1 Image-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Reference to video + + + + Try the Q1 Reference-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Start end to video + + + + Try the Q1 Start-End-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Video extension + + + + Try the Q1 Video Extension workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/vidu/vidu-q2.mdx b/tutorials/partner-nodes/vidu/vidu-q2.mdx new file mode 100644 index 000000000..d364c775c --- /dev/null +++ b/tutorials/partner-nodes/vidu/vidu-q2.mdx @@ -0,0 +1,82 @@ +--- +title: "Vidu Q2 - Fast generation with camera control" +description: "Generate video 3x faster with multi-reference support, cinematic camera controls, and enhanced dynamic rendering using Vidu Q2 in ComfyUI" +sidebarTitle: "Vidu Q2" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Vidu Q2 builds on the Q1 foundation with significantly faster inference (3x speedup), support for up to 7 reference images, enhanced subject consistency, and cinematic camera language controls including push, pull, orbit, follow, and close-up strategies. It also delivers higher-fidelity dynamic rendering with smoother large motions and more believable physical feedback, along with finer facial expressions, eye movement, and subtle gestures for more expressive characters. + +Available in Turbo, Pro, and Standard model variants. + +## Key capabilities + +- **3x faster inference** — Significantly reduced generation time compared to Q1 +- **Multi-image reference** — Up to 7 reference images ensuring identity, scene, and narrative consistency across subjects +- **Cinematic camera control** — Push, pull, orbit, follow, close-up and other camera language for precise shot direction +- **High-fidelity dynamic rendering** — Smoother large motions with more believable physical feedback +- **Expressions & micro-movements** — Finer facial expressions, eye movement, and subtle gestures for expressive characters +- **Multi-character coordination** — Stable character interaction and complex scene composition across shots +- **Six generation modes** — Text-to-video, image-to-video, reference-to-video, first-last-frame-to-video, video extension, and multi-frame generation + + + + +## Available workflows + +### Text to video + + + + Try the Q2 Text-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Image to video + + + + Try the Q2 Image-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Reference to video + + + + Try the Q2 Reference-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### First last frame to video + + + + Try the Q2 First-Last-Frame-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### First last frame to video + + + + Try the Q2 First-Last-Frame-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + diff --git a/tutorials/partner-nodes/vidu/vidu-q3.mdx b/tutorials/partner-nodes/vidu/vidu-q3.mdx new file mode 100644 index 000000000..2ae672b0b --- /dev/null +++ b/tutorials/partner-nodes/vidu/vidu-q3.mdx @@ -0,0 +1,51 @@ +--- +title: "Vidu Q3 - Latest generation for narrative and animation" +description: "Generate high-quality video with improved narrative understanding, outstanding anime output, and audio support using Vidu Q3 in ComfyUI" +sidebarTitle: "Vidu Q3" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +Vidu Q3 is the newest and most capable Vidu model generation, powered by the Vidu 3.5 engine. It delivers the highest visual quality with a particular strength in anime and 2D animation styles, making it the top choice for animation production pipelines. Q3 also supports synchronized audio generation including dialogue, sound effects, and background music. + +Available in **Pro** (up to 2K resolution) and **Turbo** (optimized for speed) tiers. + +## Key capabilities + +- **Vidu 3.5 engine** — The latest engine delivering the highest visual quality across the Vidu family +- **Exceptional anime quality** — Industry-leading output for 2D animation production +- **Synchronized audio** — Generate dialogue, sound effects, and background music matching the scene +- **Multi-resolution support** — 720p, 1080p on all models, plus 2K on Pro for image-to-video +- **Flexible duration** — Create videos from 1 to 16 seconds +- **Motion amplitude control** — Adjust movement intensity from subtle to dynamic +- **Three generation modes** — Text-to-video, image-to-video, and start/end-frame-to-video + + + + +## Available workflows + +### Text to video + + + + Try the Q3 Text-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + +### Image to video + + + + Try the Q3 Image-to-Video workflow instantly on Comfy Cloud. + + + Download the workflow JSON. + + + + diff --git a/tutorials/partner-nodes/wavespeed/overview.mdx b/tutorials/partner-nodes/wavespeed/overview.mdx new file mode 100644 index 000000000..272e584ef --- /dev/null +++ b/tutorials/partner-nodes/wavespeed/overview.mdx @@ -0,0 +1,35 @@ +--- +title: "Image & Video Enhancement with WaveSpeed" +description: "Upscale video and images, and restore old photos using WaveSpeed AI models — FLSH-VSR video upscaling, SeedVR2 image restoration, and general image upscaling, all within ComfyUI." +sidebarTitle: "WaveSpeed" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +WaveSpeed is a high-performance AI media enhancement platform, now available as native workflow nodes in ComfyUI. It offers **image upscaling** and **SeedVR2 AI image fix** for photos, plus **FLSH-VSR video upscaling** — all powered by ComfyCloud with no local GPU needed. + +## Key capabilities + +- **FLSH-VSR video upscaling** — Temporal-coherent AI video super-resolution that maintains consistent detail across frames, ideal for archiving, AI video enhancement, and SD-to-HD conversion. +- **General image upscaling** — Multi-purpose image enlargement and quality improvement covering photographs, digital art, renders, and screenshots. +- **SeedVR2 AI image fix** — Diffusion-based restoration model for recovering damaged, degraded, or low-quality images, including old photo repair, noise and artifact removal, and lost detail reconstruction. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure, keeping your local hardware free for generation and composition. +- **Easy pipeline integration** — Simple drop-in nodes that work as a final enhancement step or intermediate preprocessing stage in any ComfyUI workflow. +- **Broad content compatibility** — Models handle diverse input types including photographs, AI art, video frames, and compressed web images. + + + + +## Models + +WaveSpeed offers two model groups in ComfyUI. The **Upscale** group covers general image upscaling and SeedVR2 restoration, while **FlashVSR** provides temporal-aware video super-resolution. + + + + General image upscaling and SeedVR2 AI image restoration. + + + Temporal-coherent AI video upscaling. + + diff --git a/tutorials/partner-nodes/wavespeed/wavespeed-flashvsr.mdx b/tutorials/partner-nodes/wavespeed/wavespeed-flashvsr.mdx new file mode 100644 index 000000000..e940431ef --- /dev/null +++ b/tutorials/partner-nodes/wavespeed/wavespeed-flashvsr.mdx @@ -0,0 +1,27 @@ +--- +title: "WaveSpeed FlashVSR" +description: "Upscale low-resolution video to higher quality using WaveSpeed's temporal-aware FLSH-VSR super-resolution model in ComfyUI." +sidebarTitle: "WaveSpeed FlashVSR" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +WaveSpeed FlashVSR (FLSH-VSR) is a temporal-coherent AI video super-resolution model that upscales low-resolution video while maintaining consistent detail across frames. Ideal for archiving, AI video enhancement, and SD-to-HD conversion. + +## Key capabilities + +- **Temporal-coherent upscaling** — Maintains frame-to-frame consistency for smooth, realistic output. +- **SD-to-HD conversion** — Upscales standard-definition video to high definition with realistic detail. +- **AI video enhancement** — Ideal for archiving vintage footage and improving compressed video quality. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure. + + + + +## Available workflows + + + Launch the FLSH-VSR video upscale workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + diff --git a/tutorials/partner-nodes/wavespeed/wavespeed-upscale.mdx b/tutorials/partner-nodes/wavespeed/wavespeed-upscale.mdx new file mode 100644 index 000000000..62a7e26a9 --- /dev/null +++ b/tutorials/partner-nodes/wavespeed/wavespeed-upscale.mdx @@ -0,0 +1,40 @@ +--- +title: "WaveSpeed Upscale" +description: "Upscale images and restore old photos using WaveSpeed AI — general image upscaling and SeedVR2 AI image restoration in ComfyUI." +sidebarTitle: "WaveSpeed Upscale" +--- + +import ReqHint from "/snippets/tutorials/partner-nodes/req-hint.mdx"; +import UpdateReminder from "/snippets/tutorials/update-reminder.mdx"; + +WaveSpeed Upscale covers two image enhancement workflows: **general image upscaling** for multi-purpose enlargement and quality improvement, and **SeedVR2 AI image fix** for diffusion-based restoration of damaged, degraded, or low-quality photos. + +## Key capabilities + +- **General image upscaling** — Multi-purpose image enlargement and quality improvement for photographs, digital art, renders, and screenshots. +- **SeedVR2 AI image fix** — Diffusion-based restoration for recovering damaged, degraded, or low-quality images, including old photo repair and artifact removal. +- **No local GPU required** — All processing runs on ComfyCloud's infrastructure. +- **Broad content compatibility** — Handles photographs, AI art, video frames, and compressed web images. + + + + +## Available workflows + +### Image Upscaling + +General-purpose AI image upscaling for photographs, renders, and digital artwork. Enlarges and enhances while preserving natural detail. + + + Launch the image upscale workflow on ComfyCloud + Download the workflow JSON for local ComfyUI + + +### SeedVR2 AI Image Fix + +Restore damaged, degraded, or low-quality images using WaveSpeed's diffusion-based restoration model. Repairs noise, artifacts, color degradation, and missing facial details. + + + Launch the SeedVR2 image fix workflow on ComfyCloud + Download the workflow JSON for local ComfyUI +