The pure C++ text-to-image pipeline, driven by the OpenVINO native API for Stable Diffusion v1.5 with LMS Discrete Scheduler, supports both static and dynamic model inference. It includes advanced features like Lora integration with safetensors and OpenVINO extension for tokenizer. This demo has been tested on the Windows platform.
C++ Packages:
- CMake: Cross-platform build tool
- OpenVINO: Model inference
- Eigen3: Lora enabling
SD Preparation could be auto implemented with build_dependencies.sh
. This script provides 2 ways to install OpenVINO 2023.1.0
: conda-forge and Download archives.
cd scripts
chmod +x build_dependencies.sh
./build_dependencies.sh
...
"use conda-forge to install OpenVINO Toolkit 2023.1.0 (C++), or download from archives? (yes/no): "
Notice: Use Intel sample writeOutputBmp function instead of OpenCV for image saving.
-
Prepare a conda python env and install dependencies:
cd scripts conda create -n SD-CPP python==3.10 pip install -r requirements.txt
-
Download a huggingface SD v1.5 model like runwayml/stable-diffusion-v1-5, here another model dreamlike-anime-1.0 is used to test for the lora enabling. Ref to the official website for model downloading.
-
Model conversion from PyTorch model to OpenVINO IR via optimum-intel. Please use the script convert_model.py to convert the model into
FP16_static
orFP16_dyn
, which will be saved into the SD folder.cd scripts python -m convert_model.py -b 1 -t FP16 -sd Path_to_your_SD_model python -m convert_model.py -b 1 -t FP16 -sd Path_to_your_SD_model -dyn
Notice: Now the pipeline support batch size = 1 only, ie. static model (1,3,512,512)
Refer this blog for python pipeline, the safetensor model is loaded via src/safetensors.h. The layer name and weight are modified with Eigen Lib
and inserted into the SD model with ov::pass::MatcherPass
in the file src/lora_cpp.hpp
.
SD model dreamlike-anime-1.0 and Lora soulcard are tested in this pipeline. Here, Lora enabling only for FP16.
Download and put safetensors and model IR into the models folder.
3 steps for OpenVINO extension for tokenizer:
- The script
convert_sd_tokenizer.py
in the scripts folder could serialize the tokenizer model IR - Build OV extension:
git clone https://github.com/apaniukov/openvino_contrib/ -b tokenizer-fix-decode
Refer to PR OpenVINO custom extension ( new feature still in experiments ) - Read model with extension in the SD pipeline
Important Notes:
- Ensure you are using the same OpenVINO environment as the tokenizer extension.
- When the negative prompt is empty, use the default tokenizer without any configuration (
-e
or--useOVExtension
). - You can find the Tokenizer Model IR and the built extension file in this repository:
extensions/libuser_ov_extensions.so
models/tokenizer/
conda activate SD-CPP
mkdir build && cd build
cmake -DCMAKE_BUILD_TYPE=Release ..
make
./SD-generate -t <text> -n <negPrompt> -s <seed> --height <output image> --width <output image> -d <debugLogger> -e <useOVExtension> -r <readNPLatent> -m <modelPath> -p <precision> -l <lora.safetensors> -a <alpha> -h <help>
Usage: OV_SD_CPP [OPTION...]
-p, --posPrompt arg
Initial positive prompt for SD (default: cyberpunk cityscape like Tokyo New York with tall buildings at dusk golden hour cinematic lighting)-n, --negPrompt arg
Default is empty with space (default: )-d, --device arg
AUTO, CPU, or GPU (default: CPU)--step arg
Number of diffusion step ( default: 20)-s, --seed arg
Number of random seed to generate latent (default: 42)--num arg
Number of image output(default: 1)--height arg
Height of output image (default: 512)--width arg
Width of output image (default: 512)--log arg
Generate logging into log.txt for debug-c, --useCache
Use model caching-e, --useOVExtension
Use OpenVINO extension for tokenizer-r, --readNPLatent
Read numpy generated latents from file-m, --modelPath arg
Specify path of SD model IR (default: ../models/dreamlike-anime-1.0)-t, --type arg
Specify the type of SD model IR (FP16_static or FP16_dyn) (default: FP16_static)-l, --loraPath arg
Specify path of lora file. (*.safetensors). (default: )-a, --alpha arg
alpha for lora (default: 0.75)-h, --help
Print usage
Example:
Positive prompt: cyberpunk cityscape like Tokyo New York with tall buildings at dusk golden hour cinematic lighting
Negative prompt: (empty, here couldn't use OV tokenizer, check the issues for details)
Read the numpy latent instead of C++ std lib for the alignment with Python pipeline
- Generate image without lora
./SD-generate -r -l ""
- Generate image with soulcard lora
./SD-generate -r
- Generate the debug logging into log.txt:
./SD-generate --log
- Generate different size image with dynamic model(C++ lib generated latent):
./SD-generate -m Your_Own_Path/dreamlike-anime-1.0 -l '' -t FP16_dyn --height 448 --width 704
The performance and image quality of C++ pipeline are aligned with Python
To align the performance with Python SD pipeline, C++ pipeline will print the duration of each model inferencing only
For the diffusion part, the duration is for all the steps of Unet inferencing, which is the bottleneck
For the generation quality, be careful with the negative prompt and random latent generation. C++ random generation with MT19937 results is differ from numpy.random.randn(). Hence, please use -r, --readNPLatent for the alignment with Python(this latent file is for output image 512X512 only)
Program optimization: In addition to inference optimization, now parallel optimization with std::for_each only and add_compile_options(-O3 -march=native -Wall) with CMake
-
Download Anaconda3 and setup Conda env SD-CPP for OpenVINO with conda-forge and use the anaconda prompt terminal for CMake
-
C++ dependencies:
-
OpenVINO: To deployment without Conda: Download archives* with OpenVINO, unzip and setup environment vars with
.\setupvars.bat
-
Eigen:
1. Download from https://gitlab.com/libeigen/eigen/-/archive/3.4.0/eigen-3.4.0.zip
2. Unzip to path C:/Eigen3/eigen-3.4.0
3. Run next step's build.bat will report error: not found Eigen3Config.cmake/eigen3-config.cmake
- Create build folder for Eigen and Open VS in this path C:/Eigen3/eigen-3.4.0/build
- Open VS's developer PS terminal to do "cmake .." and redo the CMake
Ref:not found Eigen3Config.cmake/eigen3-config.cmake
- CMake with command lines, create a script build.bat:
rmdir /Q /S build
mkdir build
cd build
cmake -G "Visual Studio 16 2019" -A x64 ^
-DCMAKE_BUILD_TYPE=Release ^
..
cmake --build . --config Release
-
Put safetensors and model IR into the models folder with the following default path:
models\dreamlike-anime-1.0\FP16_static
models\soulcard.safetensors
-
Run with prompt:
cd PROJECT_SOURCE_DIR\build
.\Release\SD-generate.exe -l '' // without lora
.\Release\SD-generate.exe -l ../models/soulcard.safetensors
Notice:
* must run command line within path of build, or .exe could not find the models
* .exe is in the Release folder
- Debug within Visual Studio(open .sln file in the
build
folder)
Notice: has issue to build OpenVINO custom extension on Windows platform, so use the default tokenizer.