• Picasso AI Logo
    Logo Picasso IA
  • Home
  • AI Image
    Nano Banana 2
  • AI Video
    Veo 3.1 Fast
  • AI Chat
    Gemini 3 Pro
  • Edit Images
  • Upscale Image
  • Remove Background
  • Text to Speech
  • Effects
    NEW
  • Generations
  • Billing
  • Support
  • Account
  1. Collection
  2. Text to Image
  3. Flux Canny Pro

Control Image Structure with Flux Canny Pro

Flux Canny Pro uses edge detection to extract the structural outline of a reference image and then generates a new image that follows that exact structure. If you have ever struggled to get an AI model to respect the composition of an existing photo or sketch, this model solves that directly by locking the spatial layout in place while giving you full control over the visual style through your text prompt. The model reads the contour lines from your control image and uses them as invisible scaffolding during generation. You can take a product photograph, extract its edges, and produce a stylized illustration with the same proportions. A dedicated control slider lets you decide how strictly the output matches your prompt versus allowing more creative variation. With support for JPG and PNG output and adjustable diffusion steps, you can produce sharp, print-ready assets or fast drafts depending on your workflow. Flux Canny Pro fits naturally into design work where structure matters as much as style. A product designer can use a technical line drawing to generate realistic renders in different materials. A photographer can re-skin a portrait in a painterly style without losing the subject's pose or expression. Open the model on Picasso IA, upload your reference, type your prompt, and see the first result in about a minute.

Official

Black Forest Labs

432.8k runs

Flux Canny Pro

2024-11-21

Commercial Use

Control Image Structure with Flux Canny Pro

Table of contents

  • Overview
  • How It Works
  • Frequently Asked Questions
  • Credit Cost
  • Features
  • Use Cases
  • Examples
Get Nano Banana Pro

Overview

Flux Canny Pro is an edge-guided image generation model that uses Canny edge detection to preserve the structure and composition of a reference image while generating entirely new visual content from a text prompt. The problem it solves is concrete: when a plain text prompt gives you no guarantee that the AI will respect a specific layout, pose, or silhouette, Flux Canny Pro extracts the edge skeleton from your reference and uses it to constrain what gets generated. The result stays structurally faithful to your source while reflecting the style, colors, and subject you describe in the prompt. You can run it directly in your browser on Picasso IA, with no installation or configuration required.

How It Works

  • Upload a reference image in JPEG, PNG, GIF, or WEBP format. This image provides the structural edge map that guides the composition of the output.
  • Write a text prompt describing the visual result you want: subject, style, lighting, colors, and any other details that define the final look.
  • Set the guidance value to control how closely the output follows your prompt. Higher values produce results that match the prompt more precisely; lower values give the model more room to vary.
  • Choose the number of diffusion steps. The default of 50 balances speed with detail. Increasing it refines textures and edges at the cost of longer processing time.
  • Select your output format (JPG or PNG), then generate. Download the result and iterate with a new seed or adjusted prompt if needed.

Frequently Asked Questions

Do I need programming skills or technical knowledge to use this? No, just open Flux Canny Pro on Picasso IA, adjust the settings you want, and hit generate.

How long does generation take? Most runs finish in under a minute at the default 50-step setting. Raising the step count increases processing time but tends to produce sharper detail in complex scenes.

What makes the control image important? The control image defines the structural skeleton of the output. The cleaner and more distinct the edges in your reference, the more precisely the model can hold to that composition. Busy or low-contrast images may produce looser structural adherence.

Can I adjust how creative or varied the output is? Yes. The guidance slider controls the trade-off between prompt adherence and output diversity. You can also enable prompt upsampling, which automatically expands your prompt to encourage more creative variation in the generated result.

What output formats are supported? You can download your image as a JPG or PNG. PNG preserves lossless quality and is better for assets that need clean edges or transparency; JPG is compact and works well for most web and social media uses.

Where can I use the images I generate? The images you produce are yours to use in client projects, marketing materials, personal portfolios, or creative work. Files are delivered without watermarks.

Credit Cost

Each generation consumes 1 credit

1 credit

or 5 credits for 5 generations

Features

Everything this model can do for you

Edge-based control

Extracts contour lines from your reference image to lock the structural composition in place.

Prompt-driven styling

Shapes the visual output through natural language while the edge map handles the spatial layout.

Tunable adherence

Adjust the control slider to balance strict prompt matching against creative interpretation.

High-step precision

Increase diffusion steps up to 50 for fine detail in textures, hair, and complex surfaces.

Dual output formats

Download results as JPG for web use or PNG for designs that need a transparent-ready format.

Seed control

Set a fixed seed to reproduce the exact same image across multiple prompt iterations.

Prompt upsampling

Enable automatic prompt expansion for richer, more varied generation without rewriting your input.

Use Cases

Take a product photo and regenerate it in a different visual style while keeping the original shape and proportions intact.

Use a rough pencil sketch as a control image to produce a finished digital illustration that matches the sketch's composition.

Re-skin a portrait photo in a painterly or comic-book style without altering the subject's pose, facial structure, or expression.

Extract the edges from an architectural photo and generate multiple design variations that share the same spatial layout.

Turn a simple wireframe or layout diagram into a realistic or stylized image by describing the visual style in the prompt.

Produce consistent product renders across different surface materials by reusing the same control image with varied text prompts.

Convert a black-and-white line drawing into a colorized scene by describing the colors and lighting in the prompt.

Examples

a photo of a car on a city street
a photo of a car on a city street
16.0s
View Example
a photo of a car on a city street
a photo of a car on a city street
23.9s
View Example

Switch Category

Effects

Text To Image

Text To Image

Text To Video

Large Language Models

Large Language Models

Text To Speech

Text To Speech

Super Resolution

Super Resolution

Lipsync

AI Music Generation

AI Music Generation

Video Editing

Speech To Text

Speech To Text

AI Enhance Videos

Remove Backgrounds

Remove Backgrounds