Diffusion models revolutionize image generation by leveraging natural language to guide the creation of multimedia content. Despite significant advancements in such generative models, challenges persist in depicting detailed human-object interactions, especially regarding pose and object placement accuracy. We introduce a training-free method named Reasoning and Correcting Diffusion (ReCorD) to address these challenges. Our model couples Latent Diffusion Models with Visual Language Models to refine the generation process, ensuring precise depictions of HOIs. We propose an interaction-aware reasoning module to improve the interpretation of the interaction, along with an interaction correcting module to refine the output image for more precise HOI generation delicately. Through a meticulous process of pose selection and object positioning, ReCorD achieves superior fidelity in generated images while efficiently reducing computational requirements. We conduct comprehensive experiments on three benchmarks to demonstrate the significant progress in solving text-to-image generation tasks, showcasing ReCorD's ability to render complex interactions accurately by outperforming existing methods in HOI classification score, as well as FID and Verb CLIP-Score.
Method | HICO-DET | V-COCO | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
\(\mathcal{S}_{\mathrm{CLIP}}\uparrow\) | \(\mathcal{S}_{\mathrm{CLIP}}^{verb}\uparrow\) | PickScore \(\uparrow\) | FID \(\downarrow\) | \(\mathrm{HOI_{Full}}\uparrow\) | \(\mathrm{HOI_{Rare}}\uparrow\) | \(\mathcal{S}_{\mathrm{CLIP}}\uparrow\) | \(\mathcal{S}_{\mathrm{CLIP}}^{verb}\uparrow\) | PickScore \(\uparrow\) | FID \(\downarrow\) | HOI \(\uparrow\) | |
SD CVPR'22 | 31.74 | 21.82 | 21.50 | 51.31 | 18.78 | 10.02 | 31.10 | 21.26 | 21.26 | 77.29 | 15.85 |
A&E ACM TOG'23 | 31.63 | 21.72 | 21.33 | 46.41 | 16.57 | 8.62 | 31.21 | 21.11 | 21.11 | 70.74 | 14.52 |
LayoutLLM-T2I ACM MM'23 | 31.63 | 22.02 | 21.01 | 38.94 | 16.98 | 8.06 | 31.65 | 21.62 | 20.88 | 59.35 | 16.64 |
BoxDiff ICCV'23 | 31.42 | 21.69 | 21.22 | 45.88 | 16.33 | 8.67 | 31.06 | 21.27 | 20.96 | 68.67 | 12.34 |
InteractDiffusion CVPR'24 | 28.72 | 21.34 | 20.40 | 29.74 | 21.57 | 10.25 | 28.34 | 20.76 | 20.16 | 49.74 | 15.78 |
MultiDiffusion ICML'23 | 31.64 | 21.81 | 21.67 | 51.51 | 22.46 | 11.15 | 32.53 | 21.31 | 21.81 | 83.27 | 17.96 |
SDXL ICLR'24 | 32.06 | 22.29 | 22.68 | 40.32 | 25.85 | 14.24 | 31.76 | 21.40 | 22.54 | 75.40 | 19.02 |
LMD TMLR'24 | 28.67 | 20.11 | 20.62 | 51.37 | 9.10 | 2.65 | 29.31 | 20.29 | 20.56 | 75.68 | 10.26 |
ReCorD (ours) | 31.92 | 22.26 | 21.49 | 37.03 | 22.86 | 12.72 | 31.60 | 21.55 | 21.31 | 58.20 | 20.00 |
ReCorD\(^\dagger\) (ours) | 32.40 | 22.65 | 22.54 | 36.72 | 26.33 | 15.39 | 31.94 | 21.84 | 22.22 | 60.74 | 22.48 |
Comparison between ReCorD and existing baselines in terms of generated image quality scores in \(\mathcal{S}_{\mathrm{CLIP}}\), \(\mathcal{S}_{\mathrm{CLIP}}^{verb}\), PickScore, FID, along with HOI classification score on HICO-DET and VCOCO. Ours\(^\dagger\) represents using SDXL as backbone.
Method | \(\mathcal{S}_{\mathrm{CLIP}}\uparrow\) | \(\mathcal{S}_{\mathrm{CLIP}}^{verb}\uparrow\) | PickScore \(\uparrow\) |
---|---|---|---|
SD CVPR'22 | 30.03 | 21.39 | 20.96 |
A&E ACM TOG'23 | 29.59 | 21.65 | 20.33 |
LayoutLLM-T2I ACM MM'23 | 30.35 | 22.13 | 20.36 |
MultiDiffusion ICML'23 | 30.59 | 21.74 | 21.14 |
SDXL ICLR'24 | 30.44 | 21.86 | 21.82 |
LMD TMLR'24 | 27.27 | 20.63 | 19.94 |
ReCorD (ours) | 30.14 | 21.94 | 20.83 |
ReCorD\(^\dagger\) (ours) | 30.71 | 22.38 | 21.64 |
Comparison with SOTAs on T2I-CompBench.
@inproceedings{jianglin2024record,
title={ReCorD: Reasoning and Correcting Diffusion for HOI Generation},
author={Jiang-Lin, Jian-Yu and Huang, Kang-Yang and Lo, Ling and Huang, Yi-Ning and Lin, Terence and Wu, Jhih-Ciang and Shuai, Hong-Han and Cheng, Wen-Huang},
booktitle={Proceedings of the 32nd ACM International Conference on Multimedia},
year={2024}
}