Uni-1 is Luma’s first unified model designed for both image creation and precision editing. Unlike most image models, Uni-1 isn’t a diffusion model — it’s a decoder-only autoregressive transformer that treats text and images as a single interleaved sequence, jointly modeling time, space, and logic in one architecture. The result: a model that reasons about your prompt before generating — decomposing instructions, resolving constraints, and planning composition like a frontier LLM. It thinks first, then draws. Beyond raw output quality, what makes it powerful is control: you can generate entirely new images or surgically modify existing ones, while guiding the system with references, seeds, and structured prompts. This guide gives you a working mental model and practical workflows, from your first image to advanced multi-reference setups.Documentation Index
Fetch the complete documentation index at: https://docs.comfy.org/llms.txt
Use this file to discover all available pages before exploring further.
What makes Uni-1 different
- Two clear modes: Create Image (generate something new) and Modify Image (edit something existing)
- Up to 9 reference images, each with a defined role
- Strong control over niche and specific visual styles
- Seed support for reproducibility and controlled iteration
- Nine aspect ratios, from ultra-tall to ultra-wide
- Text rendering that’s actually readable
- Web search grounding for real-world context
- Multilingual prompts
- Multi-panel output with temporal consistency
Strengths
Uni-1 excels across a wide range of tasks:- Photorealism with material accuracy
- Illustration & stylized art with strong aesthetic control
- Old photo restoration and vintage reproduction
- Surreal and conceptual compositions
- Text rendering — readable text inside images, great for infographics and posters
- Image editing and multi-turn refinement
- Reference-guided generation with identity preservation
- Multi-panel output — consistent characters/scenes across multiple frames
The core distinction: Create vs Modify
Everything in Uni-1 starts with one question: Am I creating something new, or changing something that already exists?| Mode | What it does | When to use |
|---|---|---|
| Create Image | Produces a brand-new composition. Can be inspired by references. | ”Create a new scene in the style of this photo” |
| Modify Image | Edits a specific input image. Preserves composition and structure unless told otherwise. | ”Make this photo look like nighttime” |
- If the output should look like a version of your input, use Modify
- If it should feel inspired but new, use Create
Getting started
You can use Uni-1 in two ways:- Comfy Cloud: Preview the template and run it directly in your browser.
- Desktop: Update ComfyUI to the latest version, find the Luma UNI-1 Image node via the Node Library, and connect it to your workflow.
Image Create workflow
Run Image Create on Cloud
Try the Image Create workflow instantly on Comfy Cloud.
Download Image Create Workflow
Download JSON or search “Luma UNI-1 Image Create” in Template Library
Image Edit workflow
Run Image Edit on Cloud
Try the Image Edit workflow instantly on Comfy Cloud.
Download Image Edit Workflow
Download JSON or search “Luma UNI-1 Image Edit” in Template Library
Core parameters
| Parameter | Description |
|---|---|
| Prompt | Your primary control. Up to 6,000 characters. Be precise. |
| Aspect ratio | Controls framing, not quality. Choose based on use case. |
| Seed | Same seed + same prompt → same result. Same seed + changed prompt → controlled variation. No seed → exploration. |
| Reference images (Create) | Up to 9 images to guide different aspects. |
| Source image (Modify) | The image you are editing. Dimensions are preserved automatically. |
Working with reference images
References only work if you tell the model what they are for. Use this structure:Create mode examples
Style referenceModify mode examples
In Modify mode, clarity is everything. Always specify what to change and what must stay untouched.Prompting guidelines
Recommended lengths:- Text-to-image → 80–250 words
- Reference-guided → 100–300 words
- Modify → 30–100 words
Golden hour, 85mm lens, shallow depth of field1970s Italian giallo film poster, high-contrast color blocking
Aspect ratios
Choose based on where the image will live:| Ratio | Use case |
|---|---|
| 1:1 | Social posts |
| 9:16 | Vertical video |
| 16:9 | Widescreen |
| 3:2 / 2:3 | Photography |
| 2:1 / 3:1 | Cinematic / panoramic |
| 1:2 / 1:3 | Ultra-tall |
Seeds: control and reproducibility
Seeds turn experimentation into systems.- Fixed seed → consistency
- No seed → exploration
- Explore with no seed
- Find a strong result
- Lock the seed
- Change one variable at a time
Advanced techniques
Character consistency
- Generate a clean, front-facing reference image
- Reuse it as
IMAGE1 (CHARACTER)in every scene - Keep the label identical across prompts
Multi-reference architecture
Assign one role per image:- IMAGE1 → character
- IMAGE2 → style
- IMAGE3 → lighting
- IMAGE4 → environment
Create → Modify chain
Use Create to explore compositions, then Modify to refine details. This is one of the most powerful workflows.Iterative refinement
- Explore (no seed)
- Lock seed
- Change one variable per generation
- Document results
Troubleshooting
| Problem | Fix |
|---|---|
| References ignored | Label each one clearly |
| Modify changes too much | Explicitly state what must stay unchanged |
| Inconsistent outputs | Lock the seed |
| Prompt partially ignored | Remove conflicts or split into steps |
| Output looks like the reference image | You may be in Modify mode |
| Character inconsistency | Reuse a canonical reference image |
Quick reference
Create Image- New compositions
- Text + up to 9 references
- Descriptive prompts
- Edits existing images
- Source image + references
- Direct, surgical prompts
The golden rules
- Label every reference
- In Modify mode, always state what should not change
- Change one variable at a time when refining
- Save prompt + seed for reproducibility
- Create = new scenes, Modify = edits