M3AE-Distill: An Efficient Distilled Model for Medical Vision–Language Downstream Tasks
Multi-modal masked autoencoder (M3AE) are widely studied medical vision–language (VL) models that can be applied to various clinical tasks. However, its large parameter size poses challenges for deployment in real-world settings. Knowledge distillation (KD) has proven effective for compressing task-...
Saved in:
| Main Authors: | Xudong Liang, Jiang Xie, Mengfei Zhang, Zhuo Bi |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
MDPI AG
2025-07-01
|
| Series: | Bioengineering |
| Subjects: | |
| Online Access: | https://www.mdpi.com/2306-5354/12/7/738 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Autocorrelation Matrix Knowledge Distillation: A Task-Specific Distillation Method for BERT Models
by: Kai Zhang, et al.
Published: (2024-10-01) -
MedAlmighty: enhancing disease diagnosis with large vision model distillation
by: Yajing Ren, et al.
Published: (2025-08-01) -
Category semantic and global relation distillation for object detection
by: Yanpeng LIANG, et al.
Published: (2025-04-01) -
Automated Dead Chicken Detection in Poultry Farms Using Knowledge Distillation and Vision Transformers
by: Ridip Khanal, et al.
Published: (2024-12-01) -
Leveraging logit uncertainty for better knowledge distillation
by: Zhen Guo, et al.
Published: (2024-12-01)