AI video tools I use, plus UI/UX Pro Max (what it is and how to use it)
Carson Rodrigues / April 01, 2026
13 min read • ––– views
Hey—last time I had to leave early. Here is the fuller picture: the AI video and media stack I actually reach for, plus UI/UX Pro Max, which is less a single app and more a design-research workflow you can attach to an AI coding assistant.
UI/UX Pro Max: what it is
UI/UX Pro Max is an agent skill aimed at people building or reviewing UIs. It packages a searchable knowledge base (styles, color palettes, typography pairings, landing patterns, chart guidance, UX rules) and a small CLI that queries it so recommendations stay consistent instead of improvised.
Think of it as: given “SaaS dashboard, fintech, minimal dark”, you get back pattern, palette, type, motion, and anti-patterns grounded in the same rule set every time.
How you use it (typical workflow)
- Clarify the product — type (SaaS, portfolio, ecommerce), industry, keywords (minimal, playful, brutalist), and stack (React, Next.js, Tailwind, etc.).
- Run a design-system pass first — the skill’s workflow starts with something equivalent to a
--design-systemquery so you get one coherent system (not random one-off tips). - Drill into domains when needed — e.g.
uxfor accessibility and motion,typographyfor font pairs,landingfor hero/CTA structure,chartfor data viz. - Optional: persist — some setups can write
design-system/MASTER.md(and per-page overrides) so later sessions inherit the same tokens and rules.
What you get out of it
- Accessibility-first habits — contrast, focus rings, labels, keyboard order, touch targets.
- Interaction polish — loading states, hover vs tap, cursor affordances, error placement.
- Layout and performance — readable line length, reduced motion, avoiding layout-thrashy animation.
- Stack-aware hints — Tailwind/HTML vs React/Next/shadcn patterns where the database differentiates.
If you use Cursor (or similar), you add the skill once; when you ask for UI work, the model is nudged to run the searches and then implement, not skip straight to generic purple gradients.
AI video, audio, and image tools (categorized)
Below is the list from my notes, grouped so you can see what each layer is for. Names change fast; treat this as a map, not a ranking.
Coding agents and orchestration
- Claude (Claude Code) — planning, scripts, glue code, refactors, and tooling around a pipeline (not a renderer by itself).
Voice, speech-to-text, and audio generation
- ElevenLabs — high-quality TTS and voice tooling; common for narration and character voices.
- Deepgram — fast STT (speech-to-text) APIs; good for transcripts and caption pipelines.
- Whisper (OpenAI) — offline-capable transcription; popular in local and server workflows.
- AssemblyAI — transcription, diarization, and audio intelligence APIs.
- Play.ht — TTS with a range of voices and workflows for content.
- Murf.ai — TTS oriented toward marketing and explainer-style voiceovers.
- Resemble AI — voice cloning and branded voice workflows.
- Bark — open research-style TTS/sound generation (good for experiments).
- Soundraw, AIVA — generative music for beds and scoring (licensing models vary).
- Epidemic Sound — licensed music library (not “AI generator” in the same sense, but standard for clean clears).
Text-to-video and generative video
- Runway ML — broad gen-video and editing features; often a first stop for experiments.
- Pika Labs — text/image-to-video with a strong social/creator loop.
- Sora (OpenAI) — high-end text-to-video (availability and policy depend on OpenAI’s rollout).
- Kaiber, Genmo — stylized music-video and motion aesthetics.
- Luma AI (Dream Machine) — fast iterations on short clips from text or images.
- Stable Video Diffusion — open weights–style video generation for self-hosted stacks.
- AnimateDiff — motion modules on top of diffusion image models (often in ComfyUI).
- ComfyUI — node-based graphs for diffusion; maximum control, steeper curve.
- Automatic1111 (video extensions) — web UI ecosystem for SD; video via extensions/community nodes.
Programmatic video (React / code-first)
- Remotion — write videos as React components: compositions, sequences, and props driven by data. Strong fit when you need repeatable renders (same layout, new numbers or copy), brand-locked templates, or a pipeline that CI can run (CLI, server/Lambda, or their cloud). Pairs well with everything above: generate assets or voice elsewhere, then compose and export in Remotion; use FFmpeg-backed encoding where Remotion orchestrates it. Not a gen-AI “type a prompt, get a clip” tool—it is the programmable timeline when you already know the motion system you want.
Avatar and “talking head” products
- Synthesia, D-ID, HeyGen, Colossyan, Rephrase.ai, Tavus — template-driven presenter and lip-sync style video for corporate, sales, and L&D. Pick based on template quality, languages, API, and data handling.
Editing, post, and “make it shippable”
- CapCut — fast cuts, captions, trends; great for short-form throughput.
- Adobe Premiere Pro (Firefly AI) — pro timeline editing with Adobe’s AI assists.
- After Effects — motion graphics and compositing; pairs with Premiere for polish.
- Descript — text-first editing (edit video by editing transcript); strong for talking heads and podcasts.
- Topaz Video AI — upscaling and restoration.
- Flowframes — frame interpolation for smoother slow motion or filling gaps.
- FFmpeg — the Swiss Army knife for muxing, encoding, trimming, and automation (often the backbone under scripts).
Clipping and repurposing
- Opus Clip, VEED.io, Wisecut — auto highlights, captions, and resize for Shorts/Reels/TikTok-style distribution.
Still images and 3D (supporting video)
- Midjourney, Leonardo AI, Ideogram, Krea AI, Scenario AI — concept art, storyboards, textures, and style references.
- Meshy AI, Spline — 3D assets and interactive 3D for motion pipelines.
- Blender — full 3D: modeling, animation, compositing; pairs with AI textures or references.
How I think about stacking them
- Idea and script — Claude (or any strong LLM) plus your own taste.
- Voice and transcript — ElevenLabs or Murf for voice; Deepgram or Whisper for text alignment.
- B-roll and motion — Runway, Pika, Luma, or open ComfyUI graphs depending on budget and control.
- Data-driven or branded motion — Remotion when the video is mostly layout, typography, charts, and timing you want to version in code and re-render on demand.
- Presenter workflows — HeyGen/Synthesia-class tools when a human on camera is not the goal.
- Finish — Premiere or CapCut; FFmpeg when you need repeatable exports; Remotion when “finish” is an automated render from a React project.
Closing
UI/UX Pro Max helps you design and ship interfaces with fewer amateur tells (contrast, motion, hierarchy). The video list is the production chain I point people at when they ask how shorts, explainers, and demos get made in 2026—mix APIs, GUIs, and open tools to match your comfort with nodes and code.
If you want a follow-up, I can turn this into a minimal ComfyUI + FFmpeg skeleton, a caption pipeline (Whisper → SRT → Premiere/CapCut), or a Remotion starter (composition + props + CLI render) next.
Want to hire me as a freelancer? Let's discuss.
Drop your message and let's discuss about your project.
Chat on WhatsAppDrop in your email ID and I will get back to you.