English | δΈζ
Pixelle_video.mp4
Just input a topic, and Pixelle-Video will automatically:
- βοΈ Write video script
- π¨ Generate AI images/videos
- π£οΈ Synthesize voice narration
- π΅ Add background music
- π¬ Create video with one click
Zero threshold, zero editing experience - Make video creation as simple as typing a sentence!
- β 2026-01-26: Added the Motion Transfer pipeline β upload a reference video and an image to transfer motion.
- β 2026-01-14: Added "Digital Human" and "Image-to-Video" pipelines, multi-language TTS voices support
- β 2026-01-06: Added RunningHub 48G VRAM machine support
- β 2025-12-28: Configurable RunningHub concurrency limit, improved LLM structured data response handling
- β 2025-12-17: Added ComfyUI API Key configuration, Nano Banana model support, API template custom parameters
- β 2025-12-10: Built-in FAQ in sidebar, fixed edge-tts version to resolve TTS service instability
- β 2025-12-08: Support multiple script split modes (paragraph/line/sentence), improved template selection with direct preview
- β 2025-12-06: Fixed video generation API URL path handling with cross-platform compatibility
- β 2025-12-05: Added Windows all-in-one package download, optimized image and video analysis workflows
- β 2025-12-04: New "Custom Media" feature - upload your photos/videos with AI-powered analysis and script generation
- β 2025-11-18: Parallel processing for RunningHub, added history page, batch video task creation support
- β Fully Automatic Generation - Input a topic, automatically generate complete video
- β AI Smart Copywriting - Intelligently create narration based on topic, no need to write scripts yourself
- β AI Generated Images - Each sentence comes with beautiful AI illustrations
- β AI Generated Videos - Support AI video generation models (like WAN 2.1) to create dynamic video content
- β AI Generated Voice - Support Edge-TTS, Index-TTS and many other mainstream TTS solutions
- β Background Music - Support adding BGM to make videos more atmospheric
- β Visual Styles - Multiple templates to choose from, create unique video styles
- β Flexible Dimensions - Support portrait, landscape and other video dimensions
- β Multiple AI Models - Support GPT, Qwen, DeepSeek, Ollama and more
- β Flexible Atomic Capability Combination - Based on ComfyUI architecture, can use preset workflows or customize any capability (such as replacing image generation model with FLUX, replacing TTS with ChatTTS, etc.)
Pixelle-Video adopts a modular design, the entire video generation process is clear and concise:
From input text to final video output, the entire process is clear and simple: Script Generation β Image Planning β Frame-by-Frame Processing β Video Composition
Each step supports flexible customization, allowing you to choose different AI models, audio engines, visual styles, etc., to meet personalized creation needs.
Here are actual cases generated using Pixelle-Video, showcasing video effects with different themes and styles:
video1.mp4Korean-speaking AI Avatar |
video2.mp4Animated Cartoon Video |
video3.mp4Dancing Kitten |
default1.mp4The Scenery Along the Journey |
default2.mp4Santa ID |
default3.mp4Why Havenβt We Found Alien Civilizations Yet? |
default.mp4How to Level Up Yourself |
default.mp4Understanding Antifragility |
default.mp4Zizhi Tongjian (Comprehensive Mirror for Aid in Governance) |
default.mp4Winter Sunlight |
default.mp4Doupo Cangqiong (Battle Through the Heavens) |
default.mp4Essential Wellness Tips |
default.mp4Side Hustle Money Making |
default.mp4Insights from Zizhi Tongjian |
π‘ Tip: All these videos are fully automatically generated by AI just by inputting a topic keyword, without any video editing experience required!
No need to install Python, uv, or ffmpeg - ready to use out of the box!
π Download Windows All-in-One Package
- Download the latest Windows All-in-One Package and extract it
- Double-click
start.batto launch the Web interface - Browser will automatically open http://localhost:8501
- Configure LLM API and image generation service in "βοΈ System Configuration"
- Start generating videos!
π‘ Tip: The package includes all dependencies, no need to manually install any environment. On first use, you only need to configure API keys.
Before starting, you need to install Python package manager uv and video processing tool ffmpeg:
Please visit the uv official documentation to see the installation method for your system:
π uv Installation Guide
After installation, run uv --version in the terminal to verify successful installation.
macOS
brew install ffmpegUbuntu / Debian
sudo apt update
sudo apt install ffmpegWindows
- Download URL: https://ffmpeg.org/download.html
- After downloading, extract and add the
bindirectory to the system environment variable PATH
After installation, run ffmpeg -version in the terminal to verify successful installation.
git clone https://github.com/AIDC-AI/Pixelle-Video.git
cd Pixelle-Video# Run with uv (recommended, will automatically install dependencies)
uv run streamlit run web/app.pyBrowser will automatically open http://localhost:8501
On first use, expand the "βοΈ System Configuration" panel and fill in:
- LLM Configuration: Select AI model (such as Qwen, GPT, etc.) and enter API Key
- Image Configuration: If you need to generate images, configure ComfyUI address or RunningHub API Key
After configuration, click "Save Configuration", and you can start generating videos!
After opening the Web interface, you will see a three-column layout. Here's a detailed explanation of each part:
Configuration is required on first use. Click to expand the "βοΈ System Configuration" panel:
Used for generating video scripts.
Quick Select Preset
- Select preset model from dropdown menu (Qwen, GPT-4o, DeepSeek, etc.)
- After selection, base_url and model will be automatically filled
- Click "π Get API Key" link to register and obtain key
Manual Configuration
- API Key: Enter your key
- Base URL: API address
- Model: Model name
Used for generating video images.
Local Deployment (Recommended)
- ComfyUI URL: Local ComfyUI service address (default http://127.0.0.1:8188)
- Click "Test Connection" to confirm service is available
Cloud Deployment
- RunningHub API Key: Cloud image generation service key
After configuration, click "Save Configuration".
- AI Generated Content: Input topic, AI automatically creates script
- Suitable for: Want to quickly generate video, let AI write script
- Example: "Why develop a reading habit"
- Fixed Script Content: Directly input complete script, skip AI creation
- Suitable for: Already have ready-made script, directly generate video
- No BGM: Pure voice narration
- Built-in Music: Select preset background music (such as default.mp3)
- Custom Music: Put your music files (MP3/WAV, etc.) in the
bgm/folder - Click "Preview BGM" to preview music
- Select TTS workflow from dropdown menu (supports Edge-TTS, Index-TTS, etc.)
- System will automatically scan TTS workflows in the
workflows/folder - If you know ComfyUI, you can customize TTS workflows
- Upload reference audio file for voice cloning (supports MP3/WAV/FLAC and other formats)
- Suitable for TTS workflows that support voice cloning (such as Index-TTS)
- Can listen directly after upload
- Enter test text, click "Preview Voice" to listen to the effect
- Supports using reference audio for preview
Determine what style of images AI generates.
ComfyUI Workflow
- Select image generation workflow from dropdown menu
- Supports local deployment (selfhost) and cloud (RunningHub) workflows
- Default uses
image_flux.json - If you know ComfyUI, you can put your own workflows in the
workflows/folder
Image Dimensions
- Set width and height of generated images (unit: pixels)
- Default 1024x1024, can be adjusted as needed
- Note: Different models have different dimension limitations
Prompt Prefix
- Controls overall image style (language needs to be English)
- Example: Minimalist black-and-white matchstick figure style illustration, clean lines, simple sketch style
- Click "Preview Style" to test effect
Determines video layout and design.
Template Naming Convention
static_*.html: Static templates (no AI-generated media, text-only styles)image_*.html: Image templates (uses AI-generated images as background)video_*.html: Video templates (uses AI-generated videos as background)
Usage
- Select template from dropdown menu, displayed grouped by dimension (portrait/landscape/square)
- Click "Preview Template" to test effect with custom parameters
- If you know HTML, you can create your own templates in the
templates/folder - π View All Template Previews
- After configuring all parameters, click "π¬ Generate Video"
- Shows real-time progress (generating script β generating images β synthesizing voice β composing video)
- Automatically shows video preview after completion
- Shows current step in real-time
- Example: "Frame 3/5 - Generating Image"
- Automatically plays after generation
- Shows video duration, file size, number of frames, etc.
- Video files are saved in the
output/folder
Q: How long does it take to use for the first time?
A: Generation time depends on the number of video frames, network conditions, and AI inference speed, typically completed within a few minutes.
Q: What if I'm not satisfied with the video?
A: You can try:
- Change LLM model (different models have different script styles)
- Adjust image dimensions and prompt prefix (change image style)
- Change TTS workflow or upload reference audio (change voice effect)
- Try different video templates and dimensions
Q: What about the cost?
A: This project fully supports free operation!
- Completely Free Solution: LLM using Ollama (local) + ComfyUI local deployment = 0 cost
- Recommended Solution: LLM using Qwen (extremely low cost, highly cost-effective) + ComfyUI local deployment
- Cloud Solution: LLM using OpenAI + Image using RunningHub (higher cost but no need for local environment)
Selection Suggestion: If you have a local GPU, recommend completely free solution, otherwise recommend using Qwen (cost-effective)
Pixelle-Video design is inspired by the following excellent open-source projects:
- Pixelle-MCP - ComfyUI MCP server, allows AI assistants to directly call ComfyUI
- MoneyPrinterTurbo - Excellent video generation tool
- NarratoAI - Film commentary automation tool
- MoneyPrinterPlus - Video creation platform
- ComfyKit - ComfyUI workflow wrapper library
Thanks for the open-source spirit of these projects! π
Scan the QR codes below to join our communities for latest updates and technical support:
| Discord Community | WeChat Group |
|---|---|
- π Encountered Issues: Submit Issue
- π‘ Feature Suggestions: Submit Feature Request
- β Give a Star: If this project helps you, feel free to give a Star for support!
This project is released under the Apache License 2.0. For details, please see the LICENSE file.