Skip to content

feat: add Video main agent for AI video generation and prompt engineering#161

Merged
marcusquinn merged 1 commit intomainfrom
feature/video-main-agent
Jan 24, 2026
Merged

feat: add Video main agent for AI video generation and prompt engineering#161
marcusquinn merged 1 commit intomainfrom
feature/video-main-agent

Conversation

@marcusquinn
Copy link
Copy Markdown
Owner

@marcusquinn marcusquinn commented Jan 24, 2026

Summary

  • Adds video.md as a new primary agent (Tab-switchable in OpenCode)
  • Owns the tools/video/ subagents: video-prompt-design, remotion, higgsfield
  • Auto-discovered alphabetically by generate-opencode-agents.sh on next setup.sh run

Changes

  • New: .agent/video.md - Primary agent with subagent frontmatter, workflows, integration points
  • Updated: .agent/AGENTS.md - Added Video to progressive disclosure table
  • Updated: README.md - Added to primary agents table (15 total), alphabetical between SEO and WordPress

Architecture

  • Uses DEFAULT_TOOLS from generate-opencode-agents.sh (write, edit, bash, read, glob, grep, webfetch, task, osgrep, augment)
  • No custom MCP tools needed (guidance/prompt-crafting agent)
  • Subagents declared in frontmatter for OpenCode filtering

Summary by CodeRabbit

  • New Features

    • Added a new Video agent to the main agents list.
  • Documentation

    • Created comprehensive Video agent specifications and workflow documentation.
    • Updated main agent registry and summary tables to include the Video agent.
    • Updated main agent count from 14 to 15.

✏️ Tip: You can customize this high-level summary in your review settings.

…ring

Create dedicated video.md primary agent that owns the tools/video/
subagents (video-prompt-design, remotion, higgsfield). Auto-discovered
by setup.sh and generate-opencode-agents.sh in alphabetical order.

Includes subagent frontmatter for OpenCode Tab switching, integration
points with content/social-media/marketing agents, and workflow
documentation for prompt engineering, Remotion, and Higgsfield pipelines.
@gemini-code-assist
Copy link
Copy Markdown

Summary of Changes

Hello @marcusquinn, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request significantly expands the system's multimedia content creation capabilities by introducing a dedicated 'Video' main agent. This new agent centralizes tools and workflows for AI video generation, programmatic video creation, and advanced prompt engineering, enabling users to efficiently produce high-quality video content with consistent characters and controlled audio.

Highlights

  • New Video Agent Introduced: A new primary agent, video.md, has been added to the system, specifically designed for AI video generation and prompt engineering.
  • Integrated Subagents: The new Video agent orchestrates specialized subagents including video-prompt-design for AI video prompt engineering, remotion for programmatic video creation with React, and higgsfield for multi-model AI generation.
  • Documentation and Discovery Updates: The AGENTS.md and README.md files have been updated to include the new Video agent, ensuring it is properly listed in the progressive disclosure table and the main agents table, and is discoverable by the generate-opencode-agents.sh script.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai bot commented Jan 24, 2026

Walkthrough

This pull request introduces a new Video agent to the framework documentation. It adds agent metadata, specifications, and routing entries across three files, establishing the Video agent as a primary orchestrator for video generation workflows with three specialized subagents.

Changes

Cohort / File(s) Summary
Agent Configuration & Routing
.agent/AGENTS.md
Adds Video row to progressive disclosure and domain summary tables with links to video-prompt-design, remotion, and higgsfield subagents
Agent Specification
.agent/video.md
New agent specification file defining metadata, capabilities, workflows, and integration points for multi-component AI video production (prompt engineering, programmatic animation, generation pipelines)
Main Documentation
README.md
Adds Video entry to main agents table with MCP configuration; updates main agent count from 14 to 15

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~5 minutes

Possibly related PRs

Poem

🎬 A new agent takes the frame,
Video flows with Remotion's flame,
Prompt design meets Higgsfield's art,
Building frames from start to start! ✨

🚥 Pre-merge checks | ✅ 3
✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately and specifically describes the main change: adding a new Video main agent for AI video generation and prompt engineering, which aligns directly with the changeset content.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.


Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@sonarqubecloud
Copy link
Copy Markdown

Copy link
Copy Markdown

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces a new 'Video' main agent for AI video generation. The changes include adding the agent's definition file and updating AGENTS.md and README.md to include this new agent. The new agent file is well-structured and detailed. My review focuses on ensuring consistency and clarity. I've noted a potential point of confusion regarding the agent's mode being set to subagent despite it being a 'Main Agent'. Additionally, there's an inconsistency in the README.md where the context7 MCP is not listed for the Video agent, even though it's included in its subagents and is listed for other similar main agents. The PR description also mentions no custom MCPs are needed, which contradicts the agent's definition.

---
name: video
description: Video creation and AI generation - prompt engineering, programmatic video, generative models, editing workflows
mode: subagent
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The mode is set to subagent, but this file defines a 'Main Agent' as stated in the H1 title and the PR description. This is potentially confusing. If subagent is the correct mode for a main agent, consider adding a comment in the frontmatter to clarify this for future maintainability. If it's incorrect, it should be updated to the proper mode.

| Research | `research.md` | Research and analysis tasks | context7, augment |
| Sales | `sales.md` | Sales operations and CRM | augment |
| SEO | `seo.md` | SEO optimization, Search Console, keyword research | gsc, ahrefs, dataforseo, serper, context7, augment |
| Video | `video.md` | AI video generation, prompt engineering, programmatic video | augment |
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

For consistency with other main agents and the subagents listed in .agent/video.md, you should also include context7 in the list of enabled MCPs. The video.md file explicitly lists context7 as a subagent for research purposes, and other main agents like Plan+, SEO, and Research also list context7 in this table.

Suggested change
| Video | `video.md` | AI video generation, prompt engineering, programmatic video | augment |
| Video | `video.md` | AI video generation, prompt engineering, programmatic video | augment, context7 |

@github-actions
Copy link
Copy Markdown
Contributor

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 406 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Sat Jan 24 02:29:57 UTC 2026: Code review monitoring started
Sat Jan 24 02:29:57 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 406
Sat Jan 24 02:29:57 UTC 2026: Qlty - 0 issues found, auto-formatting applied
Sat Jan 24 02:29:59 UTC 2026: Codacy analysis completed with auto-fixes

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 406
  • VULNERABILITIES: 0

Generated on: Sat Jan 24 02:31:07 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

@augmentcode
Copy link
Copy Markdown

augmentcode bot commented Jan 24, 2026

🤖 Augment PR Summary

Summary: Adds a new video main agent focused on AI video generation, prompt engineering, and programmatic video workflows.

Changes:

  • Introduces .agent/video.md with subagent routing to video prompt design, Remotion, and Higgsfield (plus research/content helpers).
  • Updates .agent/AGENTS.md to include a Video domain entry in the progressive disclosure table.
  • Updates README.md to list Video in the main agents table and bumps the total from 14 → 15.

🤖 Was this summary useful? React with 👍 or 👎

Copy link
Copy Markdown

@augmentcode augmentcode bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Review completed. No suggestions at this time.

Comment augment review to trigger a new review at any time.

Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

🤖 Fix all issues with AI agents
In @.agent/video.md:
- Around line 21-83: Add a short credential storage/security note to the Quick
Reference section of .agent/video.md: mention the canonical credential docs in
AGENTS.md and the local helper script setup-local-api-keys.sh, instruct users to
store API keys in ~/.config/aidevops/mcp-env.sh with 600 permissions and never
in working directories, and reference the related auth docs in higgsfield.md for
completeness so readers know where to securely place keys for Veo 3, Sora,
Kling, Seedance, Higgsfield and Remotion.
🧹 Nitpick comments (1)
.agent/video.md (1)

54-77: Consider progressive disclosure: move detailed workflows to subagents.

The three workflow sections contain step-by-step implementation details that would be better suited for the respective subagent files (video-prompt-design.md, remotion.md, higgsfield.md). The main agent file should provide high-level workflow descriptions and pointer references to the subagents for detailed guidance.

This aligns with the framework's progressive disclosure pattern, keeping the main agent focused and reducing token usage until specific workflows are needed.

♻️ Recommended refactor to align with progressive disclosure pattern
 ## Workflows
 
-### AI Video Prompt Engineering
-
-1. Define character with 15+ attributes for consistency
-2. Structure prompt using 7 components (Subject, Action, Scene, Style, Dialogue, Sounds, Technical)
-3. Include camera positioning syntax and negative prompts
-4. Specify environment audio explicitly to prevent hallucinations
-5. Keep dialogue to 12-15 words for 8-second generations
-
-### Programmatic Video (Remotion)
-
-1. Define compositions with `useCurrentFrame()` and `useVideoConfig()`
-2. Drive all animations via `interpolate()` or `spring()`
-3. Use `<Sequence>` for time-offset content
-4. Render via CLI or Lambda for production
-
-### AI Generation Pipeline (Higgsfield)
-
-1. Generate base image with text-to-image (Soul, FLUX)
-2. Create character for consistency across generations
-3. Convert to video with image-to-video (DOP, Kling, Seedance)
-4. Poll for completion via webhooks or status API
+### AI Video Prompt Engineering
+
+Use the 7-component meta prompt framework for consistent AI video generation. See `@video-prompt-design` for character definition, camera positioning, audio design, and hallucination prevention strategies.
+
+### Programmatic Video (Remotion)
+
+Create videos programmatically using React components with frame-based animations. See `@remotion` for composition patterns, animation APIs, sequence management, and rendering workflows.
+
+### AI Generation Pipeline (Higgsfield)
+
+Access 100+ generative media models through a unified API. See `@higgsfield` for text-to-image, image-to-video workflows, character consistency, and webhook integration.

This keeps the main agent file concise while directing users to the appropriate subagents for implementation details.

Based on coding guidelines: Use progressive disclosure pattern with pointers to subagents instead of inline content.

Comment on lines +21 to +83
# Video - Main Agent

<!-- AI-CONTEXT-START -->

## Quick Reference

- **Purpose**: AI video generation and programmatic video creation
- **Subagents**: `tools/video/` (prompt design, Remotion, Higgsfield)

**Capabilities**:
- AI video prompt engineering (Veo 3, Sora, Kling, Seedance)
- Programmatic video creation with React (Remotion)
- Multi-model AI generation via unified API (Higgsfield)
- Character consistency across video series
- Audio design and hallucination prevention

**Typical Tasks**:
- Craft structured prompts for AI video generation
- Build programmatic video pipelines
- Generate consistent character series
- Design camera work, dialogue, and audio
- Compare and select AI video models

<!-- AI-CONTEXT-END -->

## Subagent Reference

| Subagent | Purpose |
|----------|---------|
| `video-prompt-design` | 7-component meta prompt framework for Veo 3 and similar models |
| `remotion` | Programmatic video creation with React - animations, compositions, rendering |
| `higgsfield` | Unified API for 100+ generative media models (image, video, voice, audio) |

## Workflows

### AI Video Prompt Engineering

1. Define character with 15+ attributes for consistency
2. Structure prompt using 7 components (Subject, Action, Scene, Style, Dialogue, Sounds, Technical)
3. Include camera positioning syntax and negative prompts
4. Specify environment audio explicitly to prevent hallucinations
5. Keep dialogue to 12-15 words for 8-second generations

### Programmatic Video (Remotion)

1. Define compositions with `useCurrentFrame()` and `useVideoConfig()`
2. Drive all animations via `interpolate()` or `spring()`
3. Use `<Sequence>` for time-offset content
4. Render via CLI or Lambda for production

### AI Generation Pipeline (Higgsfield)

1. Generate base image with text-to-image (Soul, FLUX)
2. Create character for consistency across generations
3. Convert to video with image-to-video (DOP, Kling, Seedance)
4. Poll for completion via webhooks or status API

## Integration Points

- `content.md` - Script writing and content planning
- `social-media.md` - Platform-specific video formatting
- `marketing.md` - Campaign video production
- `seo.md` - Video SEO (titles, descriptions, thumbnails)
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

🧩 Analysis chain

🏁 Script executed:

# Check the directory structure for video-related agents
find .agent -type f -name "*.md" | grep -i video | head -20

Repository: marcusquinn/aidevops

Length of output: 898


🏁 Script executed:

# Search for credential-related content in video-related files
rg -i "api.{0,5}key|credential|auth|token|secret" .agent --glob "*.md" | grep -i video

Repository: marcusquinn/aidevops

Length of output: 4110


🏁 Script executed:

# Check main AGENTS.md for security protocols and credential guidance
cat -n .agent/AGENTS.md | head -100

Repository: marcusquinn/aidevops

Length of output: 4167


🏁 Script executed:

# Look for the setup script mentioned in learnings
find . -name "setup-local-api-keys.sh" -o -path "*scripts/setup*" | head -20

Repository: marcusquinn/aidevops

Length of output: 188


🏁 Script executed:

# Check the full content of video.md
cat -n .agent/video.md | head -150

Repository: marcusquinn/aidevops

Length of output: 3234


🏁 Script executed:

# Check if higgsfield.md authentication section is comprehensive
cat -n .agent/tools/video/higgsfield.md | grep -A 20 "## Authentication"

Repository: marcusquinn/aidevops

Length of output: 1350


🏁 Script executed:

# Verify the setup script exists and what it does
head -50 .agent/scripts/setup-local-api-keys.sh

Repository: marcusquinn/aidevops

Length of output: 1264


Add credential storage reference to video.md security context.

The video.md agent references multiple third-party services (Veo 3, Sora, Kling, Seedance, Higgsfield, Remotion) that require API keys. While higgsfield.md documents authentication, video.md should reference the main AGENTS.md credential storage protocols or the setup-local-api-keys.sh script in its Quick Reference section to ensure users understand where to securely store these keys (~/.config/aidevops/mcp-env.sh with 600 permissions, never in working directories).

🤖 Prompt for AI Agents
In @.agent/video.md around lines 21 - 83, Add a short credential
storage/security note to the Quick Reference section of .agent/video.md: mention
the canonical credential docs in AGENTS.md and the local helper script
setup-local-api-keys.sh, instruct users to store API keys in
~/.config/aidevops/mcp-env.sh with 600 permissions and never in working
directories, and reference the related auth docs in higgsfield.md for
completeness so readers know where to securely place keys for Veo 3, Sora,
Kling, Seedance, Higgsfield and Remotion.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant