Models>

DPO (Direct Preference Optimization) LoRA for XL and 1.5 - OpenRail++

RAW photo, a close-up picture of a dog, blue eyes, reflection in it's eyes
RAW photo, a close-up picture of a cat, orange eyes, reflection in it's eyes
Portrait of a young man with a chiseled jawline, emerald green eyes, and tousled brown hair. He wears a Greek warrior's armor and stands on a cliff overlooking the Aegean Sea. he is holding a sign saying "AIEasyPic"
Quick Generate
Select a style, write what you want to see and click Generate
What you want in the image.
Clear
Random
* This model is not available for use and might generate different characters compared to the ones seen in the images.

Model Description by Creator

What is DPO?

DPO is Direct Preference Optimization, the name given to the process whereby a diffusion model is finetuned based on human-chosen images. Meihua Dang et. al. have trained Stable Diffusion 1.5 and Stable Diffusion XL using this method and the Pick-a-Pic v2 Dataset, which can be found at https://huggingface.co/datasets/yuvalkirstain/pickapic_v2, and wrote a paper about it at https://huggingface.co/papers/2311.12908.

What does it Do?

The trained DPO models have been observed to produce higher quality images than their untuned counterparts, with a significant emphasis on the adherence of the model to your prompt. These LoRA can bring that prompt adherence to other fine-tuned Stable Diffusion models.

Who Trained This?

These LoRA are based on the works of Meihua Dang (https://huggingface.co/mhdang) at

https://huggingface.co/mhdang/dpo-sdxl-text2image-v1 and https://huggingface.co/mhdang/dpo-sd1.5-text2image-v1, licensed under OpenRail++.

How were these LoRA Made?

They were created using Kohya SS by extracting them from other OpenRail++ licensed checkpoints on CivitAI and HuggingFace.

1.5: https://civitai.com/models/240850/sd15-direct-preference-optimization-dpo extracted from https://huggingface.co/fp16-guy/Stable-Diffusion-v1-5_fp16_cleaned/blob/main/sd_1.5.safetensors.

XL: https://civitai.com/models/238319/sd-xl-dpo-finetune-direct-preference-optimization extracted from https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/sd_xl_base_1.0_0.9vae.safetensors

These are also hosted on HuggingFace at https://huggingface.co/benjamin-paine/sd-dpo-offsets/

Images Generated With This Model

Loading...

Train Models

Train AI Models On Your Own Pictures

You can use your own pictures to train AI models. Reimagine yourself in new ways with our AI-powered image generator.
Create pictures of yourself around the world, in your dream outfit, as a video game character, or in countless art styles. The possibilities are endless.
Train Your Own Model

Generate AI Images. Simple and Easy

AIEasyPic is the #1 AI image generator. Generate images with simple prompts, swap faces, train models to reimagine yourself, and more.