Tuning Diffusion Models Is Simpler Than You Think

Author: Denis Avetisyan


New research demonstrates a remarkably data-efficient method for aligning powerful generative models with human preferences using a streamlined fine-tuning process.

CRAFT demonstrates an efficient fine-tuning method for diffusion models, enabling enhanced adherence to complex instructions and compositional reasoning-specifically excelling at stylistic nuance, accurate object and attribute rendering, and precise on-image text generation-compared to the base Vanilla-SDXL model.
CRAFT demonstrates an efficient fine-tuning method for diffusion models, enabling enhanced adherence to complex instructions and compositional reasoning-specifically excelling at stylistic nuance, accurate object and attribute rendering, and precise on-image text generation-compared to the base Vanilla-SDXL model.

The CRAFT framework achieves state-of-the-art alignment with only 100 samples by combining composite reward filtering and supervised fine-tuning of diffusion models.

Despite recent advances in aligning diffusion models with human preferences, current techniques remain hampered by substantial data and computational demands. This work, ‘CRAFT: Aligning Diffusion Models with Fine-Tuning Is Easier Than You Think’, introduces a novel framework that dramatically improves data efficiency and accelerates convergence during fine-tuning. By combining composite reward filtering with supervised fine-tuning, CRAFT achieves state-of-the-art performance with as few as 100 samples, and theoretically optimizes a lower bound of group-based reinforcement learning. Could this lightweight paradigm unlock broader accessibility and application of preference-aligned generative models?


The Illusion of Alignment: Chasing Subjectivity in AI Art

The remarkable capabilities of diffusion models, such as Stable Diffusion, in generating photorealistic and imaginative imagery are increasingly apparent, yet a critical challenge persists: aligning these outputs with the complex and often subjective preferences of human observers. While these models excel at learning the statistical patterns within vast datasets of images, they often struggle to capture the subtle nuances of aesthetic quality, leading to outputs that, though technically proficient, may lack artistic merit or fail to resonate with human sensibilities. This misalignment isn’t merely a matter of personal taste; it extends to concerns regarding harmful biases, stereotypical representations, and the generation of content that deviates from desired ethical standards. Consequently, significant research is now focused on developing techniques that can effectively bridge the gap between algorithmic generation and human expectation, ensuring that these powerful tools produce not only visually impressive but also genuinely desirable and appropriate content.

Efforts to refine generative AI through traditional supervised fine-tuning frequently encounter limitations when addressing aesthetic preferences. These methods rely on datasets where humans have labeled images according to specific qualities, but subjective concepts like ‘beauty’ or ‘creativity’ prove difficult to quantify and consistently replicate in training data. Consequently, models often produce outputs that, while technically proficient, lack originality or fail to resonate with human sensibilities, resulting in images perceived as bland, predictable, or simply undesirable. The challenge lies not in the model’s capacity to generate an image, but in its ability to generate an image that aligns with the complex and often unspoken criteria of human aesthetic judgment, a task that demands more than simple categorization.

CRAFT-SDXL fine-tuning enables the generation of visually compelling images exhibiting exceptional aesthetic quality.
CRAFT-SDXL fine-tuning enables the generation of visually compelling images exhibiting exceptional aesthetic quality.

The Feedback Loop: A Necessary Evil

Reinforcement Learning from Human Feedback (RLHF) presents a viable method for training generative models to align with human preferences; however, its success is contingent upon the development of an effective reward model. This model functions as a proxy for human evaluation, assigning a scalar value representing the quality of a generated image. Without a robust reward model capable of accurately quantifying desirable image characteristics, the reinforcement learning process lacks the necessary signal to effectively guide the generative model towards producing preferred outputs. The reward model, therefore, bridges the gap between subjective human judgment and the objective optimization criteria required by reinforcement learning algorithms.

ImageReward functions as a learned scoring function utilized in reinforcement learning pipelines for image generation. It is specifically trained on datasets of human preference comparisons – instances where human annotators indicate which of two generated images is more aesthetically pleasing or better aligned with a given prompt. This training enables ImageReward to predict a scalar reward value for any given image, representing its estimated alignment with human expectations. During the reinforcement learning process, this reward signal guides the generative model, encouraging it to produce images that maximize the predicted human preference score, thereby improving perceived quality and adherence to user intent without requiring continuous direct human evaluation.

Directly optimizing reinforcement learning models with human feedback presents significant computational challenges due to the cost and time required for human annotation at scale. Each iteration of model training necessitates new human preference data, creating a bottleneck for large datasets and complex models. This process is not only expensive in terms of monetary cost for annotators, but also in terms of processing time and infrastructure needed to manage the feedback loop. Consequently, research focuses on alternative methods such as reward modeling – training a separate model to approximate human preferences – to reduce reliance on continuous direct human input and enable more efficient scaling of reinforcement learning algorithms.

Human evaluation confirms that CRAFT outperforms baseline methods, aligning with automatic reward-based metrics.
Human evaluation confirms that CRAFT outperforms baseline methods, aligning with automatic reward-based metrics.

CRAFT: A Pragmatic Approach to Preference Alignment

CRAFT mitigates the drawbacks of direct preference optimization by integrating supervised fine-tuning with reinforcement learning. This combined approach leverages the strengths of both methods – supervised learning provides a strong initial policy, while reinforcement learning refines it based on reward signals. Crucially, CRAFT employs composite reward filtering to strategically select training samples. This filtering process prioritizes examples evaluated using multiple metrics, including HPS v2.1, PickScore, and AES, to ensure the selected samples are representative and contribute to a more stable and efficient training process. By focusing on high-quality, representative data, CRAFT aims to improve both the speed and performance of the fine-tuning process compared to methods relying on less curated datasets.

Composite Reward Filtering functions by selectively prioritizing training samples based on evaluations from multiple metrics – specifically, HPS v2.1, PickScore, and AES – to improve the stability and efficiency of the fine-tuning process. This filtering mechanism ensures that the training distribution remains consistent by focusing on samples that perform well across these diverse evaluation criteria. The use of multiple metrics mitigates the risk of overfitting to a single reward signal and contributes to a more robust and generalizable policy. By concentrating training efforts on high-quality samples as determined by these metrics, CRAFT minimizes the impact of noisy or misleading reward signals, leading to faster convergence and improved performance.

CRAFT utilizes group-based reinforcement learning to significantly accelerate policy optimization during fine-tuning. This approach achieves a 220x reduction in training time when compared to conventional reinforcement learning methods. Performance gains are demonstrated through improvements on the HPS v2.1 metric, indicating enhanced sample efficiency and faster convergence. The grouping strategy allows for more effective exploration and exploitation of the reward landscape, leading to a more streamlined training process without sacrificing model performance.

Evaluations demonstrate that the CRAFT framework achieves a GenEval Overall score of 57.97, representing a performance improvement over the base SDXL model, which scored 55.05 on the same metric. Furthermore, CRAFT attains a GenEval Single Object score of 99.06, indicating a high degree of accuracy in generating individual objects within a scene. These results establish CRAFT as a state-of-the-art solution for text-to-image generation, as measured by the GenEval benchmark.

CRAFT achieves superior image generation with robust geometric control and aesthetic quality across diverse conditions (e.g., Canny edge and depth maps) while demonstrating significantly improved data efficiency compared to existing fine-tuning methods like SDXL, Diff-DPO, MaPO, and SmPO.
CRAFT achieves superior image generation with robust geometric control and aesthetic quality across diverse conditions (e.g., Canny edge and depth maps) while demonstrating significantly improved data efficiency compared to existing fine-tuning methods like SDXL, Diff-DPO, MaPO, and SmPO.

Beyond the Hype: Scaling Preference Alignment for Practical Applications

The CRAFT framework distinguishes itself through a flexible training paradigm, adept at leveraging both established and emerging preference alignment techniques. It seamlessly integrates offline methods, such as Direct Preference Optimization (DPO), which utilize datasets of pre-existing human preferences to guide model learning. Crucially, CRAFT also supports online methods, dynamically creating preference pairs during the training process itself – allowing the model to actively solicit and incorporate feedback. This dual capability provides a significant advantage, enabling researchers and developers to choose the most appropriate approach based on data availability and computational resources, while also opening avenues for continuous learning and adaptation based on real-time user interaction.

The integration of Large Language Models represents a significant advancement in preference alignment techniques. These models move beyond simple binary preference judgements, offering contextual understanding that allows for more nuanced feedback during training. Rather than merely indicating which of two images is ‘better’, the language models can articulate why a particular image aligns more closely with desired attributes-such as aesthetic quality, adherence to a prompt, or overall creative direction. This detailed feedback enriches the learning process, enabling algorithms to not only satisfy preferences but also to generalize to novel situations and produce images that are both high-quality and demonstrably aligned with complex, human expectations. This capability fosters a more robust and adaptable system for generating content tailored to specific artistic or functional requirements.

Evaluations on the challenging Parti-Prompt dataset reveal that CRAFT achieves an impressive 85.7% win rate when directly compared to existing methods, signifying a substantial advancement in image generation technology. This high success rate isn’t merely a statistical figure; it demonstrates CRAFT’s ability to consistently produce images that align with nuanced human preferences. The framework’s scalability, evidenced by its performance across a diverse range of prompts, suggests it can be readily adapted to various creative applications. Consequently, CRAFT offers a versatile and robust solution for researchers and developers seeking to generate not only technically sound images, but also visually compelling and aesthetically pleasing content, pushing the boundaries of human-AI collaboration in creative domains.

CRAFT-SDXL demonstrably outperforms alternative preference optimization methods-including Vanilla, Diff-DPO, and SPO-in generating high-fidelity SDXL images with improved detail, composition, and text rendering.
CRAFT-SDXL demonstrably outperforms alternative preference optimization methods-including Vanilla, Diff-DPO, and SPO-in generating high-fidelity SDXL images with improved detail, composition, and text rendering.

The pursuit of perfectly aligned generative models often feels like chasing a mirage. This work on CRAFT, demonstrating preference alignment with a mere hundred samples, is less a revolution and more a pragmatic acknowledgement of production realities. As David Marr observed, “Representation is the key to understanding.” CRAFT sidesteps the need for massive datasets and complex reinforcement learning loops by focusing on a representation of preference through composite reward filtering and supervised fine-tuning. It’s an expensive way to complicate things, certainly, but a remarkably efficient one, suggesting that elegant theories often give way to expedient solutions when faced with the cold logic of deployment. The framework acknowledges that even the most sophisticated models require grounding in practical data constraints.

What’s Next?

The claim of ‘easier than you think’ always precedes a new category of difficulty. This work demonstrates a path toward preference alignment with limited data, a necessary step, but it simply shifts the bottleneck. The bug tracker will soon fill with cases of reward hacking-subtle failures in the composite reward that produce outputs technically aligned, yet deeply undesirable. Data efficiency isn’t about avoiding data; it’s about delaying the inevitable cost of labeling the corner cases.

The current framework, like all things, will prove brittle. Scaling to more complex preference structures-beyond simple ‘good’ or ‘bad’-will demand innovations beyond composite rewards. The real challenge isn’t achieving alignment with current human preferences, but anticipating their drift. Preference is a moving target, and a model aligned today is a liability tomorrow.

It is tempting to speak of ‘general’ preference alignment, but that’s a phantom goal. The cost of maintaining alignment will always exceed the initial investment. The system doesn’t deploy-it lets go. And then someone opens a ticket.


Original article: https://arxiv.org/pdf/2603.18991.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-22 04:27