Models>

DPO (Direct Preference Optimization) LoRA for XL and 1.5 - OpenRail++

Quick Generate
Select a style, write what you want to see and click Generate
What you want in the image.
Clear
Random
* This model is not available for use and might generate different characters compared to the ones seen in the images.

Model Description by Creator

What is DPO?

DPO is Direct Preference Optimization, the name given to the process whereby a diffusion model is finetuned based on human-chosen images. Meihua Dang et. al. have trained Stable Diffusion 1.5 and Stable Diffusion XL using this method and the Pick-a-Pic v2 Dataset, which can be found at https://huggingface.co/datasets/yuvalkirstain/pickapic_v2, and wrote a paper about it at https://huggingface.co/papers/2311.12908.

What does it Do?

The trained DPO models have been observed to produce higher quality images than their untuned counterparts, with a significant emphasis on the adherence of the model to your prompt. These LoRA can bring that prompt adherence to other fine-tuned Stable Diffusion models.

Who Trained This?

These LoRA are based on the works of Meihua Dang (https://huggingface.co/mhdang) at

https://huggingface.co/mhdang/dpo-sdxl-text2image-v1 and https://huggingface.co/mhdang/dpo-sd1.5-text2image-v1, licensed under OpenRail++.

How were these LoRA Made?

They were created using Kohya SS by extracting them from other OpenRail++ licensed checkpoints on CivitAI and HuggingFace.

1.5: https://civitai.com/models/240850/sd15-direct-preference-optimization-dpo extracted from https://huggingface.co/fp16-guy/Stable-Diffusion-v1-5_fp16_cleaned/blob/main/sd_1.5.safetensors.

XL: https://civitai.com/models/238319/sd-xl-dpo-finetune-direct-preference-optimization extracted from https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/sd_xl_base_1.0_0.9vae.safetensors

These are also hosted on HuggingFace at https://huggingface.co/benjamin-paine/sd-dpo-offsets/

Images Generated With This Model

Loading...

Similar Models

LCM-LoRA Weights - Stable Diffusion Acceleration Module
V1
LCM-LoRA Weights - Stable Diffusion Acceleration Module
(4.83)
0
35 Ratings
SDXL DPO-Turbo-LoRA
V2
SDXL DPO-Turbo-LoRA
(5.00)
0
1 Ratings
SDXL DPO-Turbo-LoRA
V2
SDXL DPO-Turbo-LoRA
(5.00)
0
2 Ratings
LCM-LoRA Weights - Stable Diffusion Acceleration Module
V2
LCM-LoRA Weights - Stable Diffusion Acceleration Module
(4.78)
0
27 Ratings
SDXL DPO-Turbo-LoRA
V1
SDXL DPO-Turbo-LoRA
(5.00)
0
1 Ratings
DPO (Direct Preference Optimization) LoRA for XL and 1.5 - OpenRail++
V2
DPO (Direct Preference Optimization) LoRA for XL and 1.5 - OpenRail++
(5.00)
0
13 Ratings

Train Models

Train AI Models On Your Own Pictures

You can use your own pictures to train AI models. Reimagine yourself in new ways with our AI-powered image generator.
Create pictures of yourself around the world, in your dream outfit, as a video game character, or in countless art styles. The possibilities are endless.
Train Your Own Model

Generate AI Images. Simple and Easy

AIEasyPic is the #1 AI image generator. Generate images with simple prompts, swap faces, train models to reimagine yourself, and more.