Image Generation with State of the Art Flux Diffusion Models

usmanmalik57 2 Tallied Votes 99 Views Share

In one of my previous articles, I explained how to generate stunning images for free using diffusion models and showed how to generate Stability AI's diffusion models for text-to-image generation.

Since then, the AI domain has progressed considerably, particularly in image generation. Black Forest Labs has released Flux.1 series of state-of-the-art vision models.

In this article, you will see how to use Flux.1 models for text-to-image generation and text-to-image modification. You will import Flux models from Hugging Face and generate images using Python code.

So, let's begin without ado.

Installing and Importing Required Libraries

Flux models are gated on Hugging Face, meaning you have to log into your account to access Flux models. To do so from a Python application, particularly Jupyter Notebook, you need to download the huggingface_hub module. In addition, you need to download the diffusers module from Hugging Face.

The script below downloads these two modules.


!pip install huggingface_hub
!pip install git+https://github.com/huggingface/diffusers.git

Note: To run scripts in this article, you will need Nvidia GPUs. You can use Google Colab, which provides free Nvidia GPUs.

Next, let's import the required libraries into our Python application:


from huggingface_hub import notebook_login
import torch
import matplotlib.pyplot as plt
from diffusers import FluxPipeline
from diffusers import FluxImg2ImgPipeline
from diffusers.utils import load_image

notebook_login() # you need to log into your hugging face account using access token

Text to Image Generation with Flux

Flux models have two variants: timestep-distilled (FLUX.1-schnell) and guidance-distilled (FLUX.1-dev). The timestep-distilled model requires fewer sampling steps and has a maximum sequence length of 256, while the guidance-distilled variant needs about 50 sampling steps for good-quality generation and has no limitations on max sequence length.

We will use the guidance-distilled Flux.1-dev model for text-to-image generation.

The following script creates a Hugging Face pipeline by importing the pretrained Flux.1-dev model from Hugging Face.


pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", torch_dtype=torch.bfloat16)
pipe.enable_model_cpu_offload()

Next, for image generation, you must pass the text prompt, the output image's height and width, the guidance scale, the number of interference steps, and the maximum sequence length for the input text.

The guidance_scale parameter influences how closely the generated image adheres to the prompt. Its value ranges between 0 and 20. The num_inference_steps determines the number of denoising steps, affecting the quality and generation time. A higher number of inference steps results in a higher-quality image but takes more time to generate.

The following script will generate an image of a girl standing in front of the Eiffel Tower, holding a sign that says, "Welcome to Paris."


prompt = "A little girl standing in front of eifel tower holding a sign that says welcome to Paris"
image = pipe(
    prompt,
    height=1024,
    width=1024,
    guidance_scale=3.5,
    num_inference_steps=50,
    max_sequence_length=512,
    generator=torch.Generator("cpu").manual_seed(0)
).images[0]
image.save("girl-in-paris.png")

Output:

girl-in-paris.png

From the above output, you can see that the model can generate a photo-realistic image.

Let's see another example. We will generate an image of a baby riding a line in Times Square, NY, with an elephant in the background.


prompt = "A baby riding a lion in time square new york with elephants in the background"
image = pipe(
    prompt,
    height=1024,
    width=1024,
    guidance_scale=10,
    num_inference_steps=50,
    max_sequence_length=512,
    generator=torch.Generator("cpu").manual_seed(0)
).images[0]
image.save("baby_lion_time_square.png")

Output:

baby_lion_time_square.png

The above output shows that the model could generate all the details specified in the text prompt.

Finally, I will create a simple function that generates an image given a prompt and the output image name. You can use this function to generate images in your code.


def generate_image_from_text(prompt, image_name):
  image = pipe(
    prompt,
    height=1024,
    width=1024,
    guidance_scale=3.5,
    num_inference_steps=50,
    max_sequence_length=512,
    generator=torch.Generator("cpu").manual_seed(0)
  ).images[0]
  image.save(image_name + ".png")


prompt = "A golden duck swimming in a lake with mountains and sunset view in the background"
image_name = "duck_in_lake"
generate_image_from_text(prompt, image_name)

Output:

duck_in_lake.png

As you can see from the above output, the Flux.1-dev model generates very high-quality images based on text prompts.

In the next section, you will see how to modify existing images based on text prompts.

Image Modification using Text Prompts in Flux

We will use the timestep-distilled Flux.1-schnell model for image modification.

The following script creates a Hugging Face pipeline for the Flux.1-schnell model.


device = "cuda"
mod_pipe = FluxImg2ImgPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", torch_dtype=torch.bfloat16)
mod_pipe = mod_pipe.to(device)

We will modify the following Wikipedia image of the Pyramid of Giza by adding birds, camels, and a river.

Input Image:

All_Gizah_Pyramids.jpg

Image modification is similar to image generation, except we must also pass the strength parameter to the pipeline object. The strength parameter defines the extent to which the original image will be modified.


url = "https://upload.wikimedia.org/wikipedia/commons/thumb/a/af/All_Gizah_Pyramids.jpg/1920px-All_Gizah_Pyramids.jpg"
init_image = load_image(url)

prompt = "add birds, camels, and blue river"

images = mod_pipe(
    prompt=prompt,
    image=init_image,
    num_inference_steps=50,
    strength=0.75,
    guidance_scale= 7.5,
    ).images[0]

images.save("pyramids_modified.png")

Output:

pyramids_modified.png

In the above image, a few birds and camels are added to the original image.

Let's increase the values of the strength and guidance_scale parameters to see how this affects the original image.


images = mod_pipe(
    prompt=prompt,
    image=init_image,
    num_inference_steps=50,
    strength=0.85,
    guidance_scale= 10.0,
    ).images[0]

images.save("pyramids_modified2.png")

Output:

pyramids_modified2.png

The above output shows that the original image has been modified much more extensively than the previous modification.

Finally, we will define the modify_image() function, which accepts the image URL, the prompt to modify the image, and the name for the modified image and modifies the passed image.


def modify_image(image_url, prompt, image_name):

  init_image = load_image(url)
  images = mod_pipe(
    prompt=prompt,
    image=init_image,
    num_inference_steps=50,
    strength=0.85,
    guidance_scale= 10.0,
    ).images[0]

  images.save(image_name + ".png")

prompt = "cars, horses"
url = "/content/1280px-Taj_Mahal,_Agra,_India_edit3.jpg"
name = "taj_mahal_modified"

modify_image(url, prompt, name)

Here is the input image.

Input Image:

1280px-Taj_Mahal,_Agra,_India_edit3.jpg

And here is the modified output. You can see some cars and horses added to the image.

Output:

taj_mahal_modified.png

Conclusion

Flux.1 models are the state-of-the-art image generation models. In this article, you saw how to generate and modify images with text prompts using the Flux.1 models. I encourage you to play around with strength and guidance scale parameters to generate and modify your custom images. Let me know if you like the results.

autowrecking 0 Newbie Poster

Impressive Thanks for sharing.

Be a part of the DaniWeb community

We're a friendly, industry-focused community of developers, IT pros, digital marketers, and technology enthusiasts meeting, networking, learning, and sharing knowledge.