Not every capability on Cliprise Models ships with a 3,000-word playbook on day one. That is normal for multi-provider platforms: APIs stabilize, pricing shifts, and editorial catches up. This page exists so you never stare at a landing card without context - you get where to click, which adjacent Learn article to steal prompts from, and when to escalate to a flagship comparison.
Quick answer: Read the model card for limits and billing, then borrow motion or still-language from the nearest sibling guide (start with multi-model workflows and AI video models ranked). Treat anything listed here as routing truth, not marketing fluff.
How to use this hub in three passes
- Pick modality - still vs motion vs audio vs cleanup utility.
- Match economics - latency-sensitive drafts vs premium finals (credit discipline matters more than model hype).
- Cross-check comparisons - situate the engine inside the broader market via the comparisons hub.
If two models disagree on a brief, that is expected: multi-model workflows win when you swap engines per shot, not when you force one badge to do everything.
Gemini image tiers (exploration vs finals)
- Gemini 2.5 Flash Image - fast passes, layout exploration, social crops.
- Gemini 2.5 Pro Image - higher fidelity stills when Flash leaves artifacts on faces or product edges.
- Gemini 3 Flash - newer-generation routing when your workspace exposes it; validate aspect ratios before batch jobs.
Workflow snapshot: Draft interiors or packaging variants in Flash, lock hero frames in Pro, then optionally animate winners through image-to-video workflow.
Read next: Nano Banana 2 vs Imagen vs Flux plus AI image generation guide.
GPT-style image (4o / GPT Image routes)
- GPT Image / 4o Image - natural-language edits and iterative requests such as 'change the lamp, keep the couch' layered on top of classic diffusion stacks.
Workflow snapshot: Use GPT-class routes when instruction following in plain English beats wrestling with slider stacks; fall back to Flux or Imagen when you need maximum photoreal texture control.
Read next: Best AI image generator comparison.
Seedance 1.x lanes (legacy speed tiers)
Workflow snapshot: Storyboard with reference frames if your UI supports it, keep prompts atomic (camera verb + subject + lighting), and compare outputs against Seedance 2.0 prompts so vocabulary stays modern even when the tier number is older.
Read next: Seedance vs Kling.
Wan specialists (open-ecosystem DNA)
Workflow snapshot: Pair Alibaba Wan coverage in Alibaba Wan notes with the routing discipline from AI video models ranked. Speech-to-video shines when audio is already locked (ADR, VO scratch tracks).
Audio and utility extras
- ElevenLabs Sound Effect v2 - pair with ElevenLabs dialogue guide when stacking sfx after VO.
- ElevenLabs V3 Text to Dialogue - multi-speaker clips; extend via text-to-speech guide.
- Recraft Remove Background - matte cleanup before upscale or animation; feature overview on AI background remover.
Specialist video utilities
- Sora 2 Watermark Remove - provider-specific cleanup when eligible; policy framing in no-watermark comparison.
- Z-Image - rapid low-cost stills; contextual ranking in best AI image generator comparison.
Before you burn credits: model card preflight
Every route listed on Cliprise Models carries constraints that hero marketing lines cannot replace. Treat the live card as the contract: resolution ceilings, duration caps, frame-rate defaults, reference uploads, optional audio coupling, and credit estimates drift whenever providers ship patches. Walk this checklist once per campaign template so reviewers stop blaming "the wrong model" when the brief violated something spelled out on the card.
Inputs and aspect ratio. Confirm whether the lane expects text-only prompts, still references, multi-angle character sheets, masks, or timed dialogue. If your hero master is 4:5 social but the motion lane defaults to widescreen cinematic, decide where you crop before you animate—retrofitting aspect after motion burns generations.
Temporal and audio truth. Note maximum clip length, interpolation behavior, and whether the workflow respects locked seeds or fixed VO clocks. Speech-aligned lanes only perform when narration timing is frozen; vague text prompts rarely fix waveform misalignment after the fact.
Economics versus fidelity. Snapshot the credit curve at the duration and resolution you actually ship. Exploration belongs on fast tiers; finals belong on premium tiers when skin texture, typography, or motion continuity fails cheaper passes. If the cost gap is wide, batch drafts in the economical lane and promote only winners—the same multi-engine discipline described in multi-model workflows, applied inside one modality.
Handoff QA. When stills feed video, inspect edges, hands, small type, and logos at full pixel scale before you queue motion. When video feeds finishing audio or sfx, leave headroom for dynamics—not every export normalizes loudness for broadcast or platform specs.
If any checkpoint conflicts with stakeholder promises, pause: borrow negotiation language from AI video models ranked or the closest comparison guide, then reset timelines instead of stacking hidden rework.
Swap the engine or regenerate in place?
Swap engines when failures cluster by modality—every Flash attempt breaks texture on fabric but Pro clears it, or one video backbone adds motion blur artifacts another avoids at similar settings. Regenerate in place when the failure looks prompt-level (missing geometry, conflicting lighting directives, absent negatives) or like ordinary stochastic noise inside an otherwise sound lane. Logging that distinction keeps credit spreadsheets honest and stops teams from mythologizing "bad models" when the brief never carried measurable constraints.
When this hub stops being enough
Escalate to a dedicated guide or a paid benchmark sprint when:
- Legal teams require documented defaults (real estate, pharma, finance overlays).
- You need reproducible seeds or LUT-matched color across dozens of shots.
- You are standardizing a single house style across regions.
Until then, bookmark this page, keep Cliprise pricing open for credit math, and branch through the comparisons hub whenever SERP research sends you conflicting claims.
When a model graduates to its own deep guide, editors trim the matching section here so the index stays honest.
