Open Access iconOpen Access

ARTICLE

Image Style Transfer for Exhibition Hall Design Based on Multimodal Semantic-Enhanced Algorithm

Qing Xie*, Ruiyun Yu

Software College, Northeastern University, Shenyang, 110000, China

* Corresponding Author: Qing Xie. Email: email

Computers, Materials & Continua 2025, 84(1), 1123-1144. https://doi.org/10.32604/cmc.2025.062712

Abstract

Although existing style transfer techniques have made significant progress in the field of image generation, there are still some challenges in the field of exhibition hall design. The existing style transfer methods mainly focus on the transformation of single dimensional features, but ignore the deep integration of content and style features in exhibition hall design. In addition, existing methods are deficient in detail retention, especially in accurately capturing and reproducing local textures and details while preserving the content image structure. In addition, point-based attention mechanisms tend to ignore the complexity and diversity of image features in multi-dimensional space, resulting in alignment problems between features in different semantic areas, resulting in inconsistent stylistic features in content areas. In this context, this paper proposes a semantic-enhanced multimodal style transfer algorithm tailored for exhibition hall design. The proposed approach leverages a multimodal encoder architecture to integrate information from text, source images, and style images, using separate encoder modules for each modality to capture shallow, deep, and semantic features. A novel Style Transfer Convolution (STConv) convolutional kernel, based on the Visual Geometry Group (VGG) 19 network, is introduced to improve feature extraction in style transfer. Additionally, an enhanced Transformer encoder is incorporated to capture contextual semantic information within images, while the CLIP model is employed for text data processing. A hybrid attention module is designed to precisely capture style features, achieving multimodal feature fusion via a diffusion model that generates exhibition hall design images aligned with stylistic requirements. Quantitative experiments show that compared with the most advanced algorithms, the proposed method has achieved significant performance improvement on both Fréchet Inception Distance (FID) and Kernel Inception Distance (KID) indexes. For example, on the ExpoArchive dataset, the proposed method has a FID value of 87.9 and a KID value of 1.98, which is significantly superior to other methods.

Keywords

Exhibition hall design; style transfer; multimodal fusion; semantic enhancement; diffusion model

Cite This Article

APA Style
Xie, Q., Yu, R. (2025). Image Style Transfer for Exhibition Hall Design Based on Multimodal Semantic-Enhanced Algorithm. Computers, Materials & Continua, 84(1), 1123–1144. https://doi.org/10.32604/cmc.2025.062712
Vancouver Style
Xie Q, Yu R. Image Style Transfer for Exhibition Hall Design Based on Multimodal Semantic-Enhanced Algorithm. Comput Mater Contin. 2025;84(1):1123–1144. https://doi.org/10.32604/cmc.2025.062712
IEEE Style
Q. Xie and R. Yu, “Image Style Transfer for Exhibition Hall Design Based on Multimodal Semantic-Enhanced Algorithm,” Comput. Mater. Contin., vol. 84, no. 1, pp. 1123–1144, 2025. https://doi.org/10.32604/cmc.2025.062712



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 320

    View

  • 126

    Download

  • 0

    Like

Share Link