In AI-empowered poster design, content-aware layout generation is crucial for the on-image arrangement of visual-textual elements, e.g., logo, text, and underlay. To perceive background images, existing work demanded a high parameter count that far exceeds the size of available training data, which has impeded the model's real-time performance and generalization ability. To address these challenges, we proposed a patch-level data summarization and augmentation approach, vividly named Scan-and-Print. Specifically, the scan procedure selects only the patches suitable for placing element vertices to perform fine-grained perception efficiently. Then, the print procedure mixes up patches and vertices across two image-layout pairs to synthesize over 100% new samples in each epoch while preserving their plausibility. To facilitate the vertex-level operations, we also introduced a vertex-based layout representation. Extensive experimental results on widely used benchmarks demonstrated that Scan-and-Print can generate visually appealing layouts with state-of-the-art quality while dramatically reducing computational bottleneck by 95.2%†.
†: Compared to the current state-of-the-art method, RALF [Horita et al., CVPR 2024 Oral].
@inproceedings{Hsu-IJCAI2025-ScanandPrint,
title={Scan-and-Print: Patch-level Data Summarization and Augmentation for Content-aware Layout Generation in Poster Design},
author={Hsu, HsiaoYuan and Peng, Yuxin},
booktitle={Proceedings of the International Joint Conference on Artificial Intelligence},
year={2025}
}