Image Generation Using Diffusers
Convert and Optimize Model
Refer to the Model Preparation guide for detailed instructions on how to download, convert and optimize models for OpenVINO GenAI.
Run Model Using OpenVINO GenAI
OpenVINO GenAI supports the following diffusion model pipelines:
Text2ImagePipeline
for creating images from text prompts.Image2ImagePipeline
for modifying existing images based on prompts.InpaintingPipeline
for selectively replacing portions of images using masks.
See all supported image generation models.
Text2ImagePipeline
- Python
- C++
- CPU
- GPU
import openvino_genai as ov_genai
from PIL import Image
pipe = ov_genai.Text2ImagePipeline(model_path, "CPU")
image_tensor = pipe.generate(prompt)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
import openvino_genai as ov_genai
from PIL import Image
pipe = ov_genai.Text2ImagePipeline(model_path, "GPU")
image_tensor = pipe.generate(prompt)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
- CPU
- GPU
Code below requires installation of C++ compatible package. See here for additional setup details, or How to Build OpenVINO™ GenAI APP in C++ blog for full instruction.
#include "openvino/genai/image_generation/text2image_pipeline.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2];
ov::genai::Text2ImagePipeline pipe(models_path, "CPU");
ov::Tensor image = pipe.generate(prompt);
imwrite("image.bmp", image, true);
}
#include "openvino/genai/image_generation/text2image_pipeline.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2];
ov::genai::Text2ImagePipeline pipe(models_path, "GPU");
ov::Tensor image = pipe.generate(prompt);
imwrite("image.bmp", image, true);
}
Use CPU or GPU as devices without any other code change.
Image2ImagePipeline
- Python
- C++
- CPU
- GPU
import openvino_genai as ov_genai
import openvino as ov
from PIL import Image
import numpy as np
def read_image(path: str) -> ov.Tensor:
pic = Image.open(path).convert("RGB")
image_data = np.array(pic)[None]
return ov.Tensor(image_data)
input_image_data = read_image("input_image.jpg")
pipe = ov_genai.Image2ImagePipeline(model_path, "CPU")
image_tensor = pipe.generate(prompt, image=input_image_data, strength=0.8)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
import openvino_genai as ov_genai
import openvino as ov
from PIL import Image
import numpy as np
def read_image(path: str) -> ov.Tensor:
pic = Image.open(path).convert("RGB")
image_data = np.array(pic)[None]
return ov.Tensor(image_data)
input_image_data = read_image("input_image.jpg")
pipe = ov_genai.Image2ImagePipeline(model_path, "GPU")
image_tensor = pipe.generate(prompt, image=input_image_data, strength=0.8)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
- CPU
- GPU
#include "openvino/genai/image_generation/image2image_pipeline.hpp"
#include "load_image.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2], image_path = argv[3];
ov::Tensor input_image = utils::load_image(image_path);
ov::genai::Image2ImagePipeline pipe(models_path, "CPU");
ov::Tensor generated_image = pipe.generate(prompt, input_image, ov::genai::strength(0.8f));
imwrite("image.bmp", generated_image, true);
}
#include "openvino/genai/image_generation/image2image_pipeline.hpp"
#include "load_image.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2], image_path = argv[3];
ov::Tensor input_image = utils::load_image(image_path);
ov::genai::Image2ImagePipeline pipe(models_path, "GPU");
ov::Tensor generated_image = pipe.generate(prompt, input_image, ov::genai::strength(0.8f));
imwrite("image.bmp", generated_image, true);
}
InpaintingPipeline
- Python
- C++
- CPU
- GPU
import openvino_genai as ov_genai
import openvino as ov
from PIL import Image
import numpy as np
def read_image(path: str) -> ov.Tensor:
pic = Image.open(path).convert("RGB")
image_data = np.array(pic)[None]
return ov.Tensor(image_data)
input_image_data = read_image("input_image.jpg")
mask_image = read_image("mask.jpg")
pipe = ov_genai.InpaintingPipeline(model_path, "CPU")
image_tensor = pipe.generate(prompt, image=input_image_data, mask_image=mask_image)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
import openvino_genai as ov_genai
import openvino as ov
from PIL import Image
import numpy as np
def read_image(path: str) -> ov.Tensor:
pic = Image.open(path).convert("RGB")
image_data = np.array(pic)[None]
return ov.Tensor(image_data)
input_image_data = read_image("input_image.jpg")
mask_image = read_image("mask.jpg")
pipe = ov_genai.InpaintingPipeline(model_path, "GPU")
image_tensor = pipe.generate(prompt, image=input_image_data, mask_image=mask_image)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
- CPU
- GPU
#include "openvino/genai/image_generation/inpainting_pipeline.hpp"
#include "load_image.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2];
ov::Tensor input_image = utils::load_image(argv[3]);
ov::Tensor mask_image = utils::load_image(argv[4]);
ov::genai::InpaintingPipeline pipe(models_path, "CPU");
ov::Tensor generated_image = pipe.generate(prompt, input_image, mask_image);
imwrite("image.bmp", generated_image, true);
}
#include "openvino/genai/image_generation/inpainting_pipeline.hpp"
#include "load_image.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2];
ov::Tensor input_image = utils::load_image(argv[3]);
ov::Tensor mask_image = utils::load_image(argv[4]);
ov::genai::InpaintingPipeline pipe(models_path, "GPU");
ov::Tensor generated_image = pipe.generate(prompt, input_image, mask_image);
imwrite("image.bmp", generated_image, true);
}
Additional Usage Options
Use Different Generation Parameters
You can adjust several parameters to control the image generation process, including dimensions and the number of inference steps:
- Python
- C++
import openvino_genai as ov_genai
from PIL import Image
pipe = ov_genai.Text2ImagePipeline(model_path, "CPU")
image_tensor = pipe.generate(
prompt,
width=512,
height=512,
num_images_per_prompt=1,
num_inference_steps=30,
guidance_scale=7.5
)
image = Image.fromarray(image_tensor.data[0])
image.save("image.bmp")
#include "openvino/genai/image_generation/text2image_pipeline.hpp"
#include "imwrite.hpp"
int main(int argc, char* argv[]) {
const std::string models_path = argv[1], prompt = argv[2];
ov::genai::Text2ImagePipeline pipe(models_path, "CPU");
ov::Tensor image = pipe.generate(
prompt,
ov::genai::width(512),
ov::genai::height(512),
ov::genai::num_images_per_prompt(1),
ov::genai::num_inference_steps(30),
ov::genai::guidance_scale(7.5f)
);
imwrite("image.bmp", image, true);
}
width
: The width of resulting image(s).height
: The height of resulting image(s).num_images_per_prompt
: Specifies how many image variations to generate in a single request for the same prompt.num_inference_steps
: Defines denoising iteration count. Higher values increase quality and generation time, lower values generate faster with less detail.guidance_scale
: Balances prompt adherence vs. creativity. Higher values follow prompt more strictly, lower values allow more creative freedom.rng_seed
: Controls randomness for reproducible results. Same seed produces identical images across runs.
For the full list of generation parameters, refer to the API reference.
Working with LoRA Adapters
For image generation models like Stable Diffusion, LoRA adapters can modify the generation process to produce images with specific artistic styles, content types, or quality enhancements.
Refer to the LoRA Adapters for more details on working with LoRA adapters.