Enhanced CLIP-GPT Framework for Cross-Lingual Remote Sensing Image Captioning
Remote Sensing Image Captioning (RSIC) aims to generate precise and informative descriptive text for remote sensing images using computational algorithms. Traditional “encoder-decoder” approaches face limitations due to their high training costs and heavy reliance on large-scal...
Saved in:
Main Authors: | Rui Song, Beigeng Zhao, Lizhi Yu |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10816156/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
KE-RSIC: Remote Sensing Image Captioning Based on Knowledge Embedding
by: Kangda Cheng, et al.
Published: (2025-01-01) -
Incidental vocabulary recognition effects of subtitled, captioned and reverse subtitled audiovisual input
by: Jana van der Kolk, et al.
Published: (2024-07-01) -
Undergraduate students’ perceptions toward writing Instagram captions in English
by: Nahda Nafisah Hutasuhut, et al.
Published: (2024-05-01) -
Combining Region-Guided Attention and Attribute Prediction for Thangka Image Captioning Method
by: Fujun Zhang, et al.
Published: (2025-01-01) -
Offline visual aid system for the blind based on image captioning
by: Yue CHEN, et al.
Published: (2022-01-01)