
![Flux.1 Kontext [pro]](https://cdn.prod.website-files.com/65b8f370a600366bc7cf9b20/6889c6d328a153aff5ccdfd0_logo.png)
FLUX.1 Kontext [pro] enables fast, consistent multi-step image editing and generation with text and image inputs.
FLUX.1 Kontext Pro is Black Forest Labs’ breakthrough in-context image model. Unlike traditional text-to-image systems that start from scratch every time, Kontext Pro understands both your text prompt and any reference images you feed it. It performs surgical local edits, full-scene transformations, and maintains perfect character, style, and identity across multiple iterations.
Built on a rectified flow transformer architecture with 12 billion parameters, it delivers FLUX-level visual quality while pioneering true iterative workflows. You can edit an image once, then edit the result again, and again without losing facial features, clothing details, lighting mood, or brand style.
It replaces clunky multi-model pipelines with a single, reliable endpoint. No more stitching together separate tools for generation, editing, and consistency checks. Just one fast, powerful call.
Compared to ByteDance BAGEL (40+ seconds) or OpenAI GPT-4o image generation, FLUX.1 Kontext Pro wins on speed, consistency, and cost every single time.
FLUX models (shown in purple) consistently achieve top performance with ELO ratings often exceeding 1100, demonstrating the strong capabilities of the Flux.1 Kontext model family in various image generation and editing benchmarks.

Feed it a reference image (or several) plus a text prompt, and watch magic happen. The model intelligently merges visual context with your instructions, perfect for “change the background to a cyberpunk city but keep the character exactly the same.”
Want to tweak just the hat on your character while leaving everything else untouched? FLUX.1 Kontext Pro does regional edits with surgical precision. No more “everything changed” disasters.
This is where Kontext Pro destroys the competition. Facial features, body type, clothing details, art style — everything stays locked in across completely different scenes and multiple editing turns. Marketers building avatar libraries or game devs creating consistent characters finally get a model that actually delivers.
Build images step by step with minimal latency. Start with a base character → add accessories → change environment → refine lighting. Each generation builds directly on the previous one. Perfect for rapid prototyping and A/B testing visuals.
Don’t need a reference image? Drop a detailed prompt and get stunning results with excellent prompt adherence, beautiful typography, and complex compositions.
Choose from 21:9, 16:9, 4:3, 1:1, 9:16 and more. Fine-tune with guidance_scale (1-20), safety_tolerance (1-6), seeds, and batch generation (up to 4 images per call).
Upload a product photo and instantly generate it in new lifestyles, colors, or environments, no expensive photoshoots required. Maintain exact branding across every variation.
Build apps that generate consistent creator avatars, YouTube thumbnails, or Instagram carousels in seconds. Your users will love the speed and visual coherence.
Create character sheets, environment concepts, and narrative scenes where the hero looks identical from frame to frame. Iterative editing makes storyboarding lightning fast.
Generate custom diagrams, historical scenes, or scientific illustrations with consistent visual language across entire lesson modules.
Teams maintain brand-consistent visuals at scale, from pitch decks to internal dashboards, without hiring extra designers.
FLUX.1 Kontext Pro is Black Forest Labs’ breakthrough in-context image model. Unlike traditional text-to-image systems that start from scratch every time, Kontext Pro understands both your text prompt and any reference images you feed it. It performs surgical local edits, full-scene transformations, and maintains perfect character, style, and identity across multiple iterations.
Built on a rectified flow transformer architecture with 12 billion parameters, it delivers FLUX-level visual quality while pioneering true iterative workflows. You can edit an image once, then edit the result again, and again without losing facial features, clothing details, lighting mood, or brand style.
It replaces clunky multi-model pipelines with a single, reliable endpoint. No more stitching together separate tools for generation, editing, and consistency checks. Just one fast, powerful call.
Compared to ByteDance BAGEL (40+ seconds) or OpenAI GPT-4o image generation, FLUX.1 Kontext Pro wins on speed, consistency, and cost every single time.
FLUX models (shown in purple) consistently achieve top performance with ELO ratings often exceeding 1100, demonstrating the strong capabilities of the Flux.1 Kontext model family in various image generation and editing benchmarks.

Feed it a reference image (or several) plus a text prompt, and watch magic happen. The model intelligently merges visual context with your instructions, perfect for “change the background to a cyberpunk city but keep the character exactly the same.”
Want to tweak just the hat on your character while leaving everything else untouched? FLUX.1 Kontext Pro does regional edits with surgical precision. No more “everything changed” disasters.
This is where Kontext Pro destroys the competition. Facial features, body type, clothing details, art style — everything stays locked in across completely different scenes and multiple editing turns. Marketers building avatar libraries or game devs creating consistent characters finally get a model that actually delivers.
Build images step by step with minimal latency. Start with a base character → add accessories → change environment → refine lighting. Each generation builds directly on the previous one. Perfect for rapid prototyping and A/B testing visuals.
Don’t need a reference image? Drop a detailed prompt and get stunning results with excellent prompt adherence, beautiful typography, and complex compositions.
Choose from 21:9, 16:9, 4:3, 1:1, 9:16 and more. Fine-tune with guidance_scale (1-20), safety_tolerance (1-6), seeds, and batch generation (up to 4 images per call).
Upload a product photo and instantly generate it in new lifestyles, colors, or environments, no expensive photoshoots required. Maintain exact branding across every variation.
Build apps that generate consistent creator avatars, YouTube thumbnails, or Instagram carousels in seconds. Your users will love the speed and visual coherence.
Create character sheets, environment concepts, and narrative scenes where the hero looks identical from frame to frame. Iterative editing makes storyboarding lightning fast.
Generate custom diagrams, historical scenes, or scientific illustrations with consistent visual language across entire lesson modules.
Teams maintain brand-consistent visuals at scale, from pitch decks to internal dashboards, without hiring extra designers.