Adaptive Multimodal Fusion with Cross-Attention for Robust Scene Segmentation and Urban Economic Analysis
With the increasing demand for accurate multimodal data analysis in complex scenarios, existing models often struggle to effectively capture and fuse information across diverse modalities, especially when data include varying scales and levels of detail. To address these challenges, this study prese...
Saved in:
Main Authors: | Chun Zhong, Shihong Zeng, Hongqiu Zhu |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2025-01-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/15/1/438 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Semantic-enhanced panoptic scene graph generation through hybrid and axial attentions
by: Xinhe Kuang, et al.
Published: (2024-12-01) -
Design of an Integrated Model for Video Summarization Using Multimodal Fusion and YOLO for Crime Scene Analysis
by: Sai Babu Veesam, et al.
Published: (2025-01-01) -
Instance-level semantic segmentation of nuclei based on multimodal structure encoding
by: Bo Guan, et al.
Published: (2025-02-01) -
Multimodal Recommendation System Based on Cross Self-Attention Fusion
by: Peishan Li, et al.
Published: (2025-01-01) -
Enhancing foundation models for scientific discovery via multimodal knowledge graph representations
by: Vanessa Lopez, et al.
Published: (2025-01-01)