Showing 1 - 20 results of 34 for search 'pre-trained transformer', query time: 0.06s Refine Results
  1. 1

    RSGPT: a generative transformer model for retrosynthesis planning pre-trained on ten billion datapoints by Yafeng Deng, Xinda Zhao, Hanyu Sun, Yu Chen, Xiaorui Wang, Xi Xue, Liangning Li, Jianfei Song, Chang-Yu Hsieh, Tingjun Hou, Xiandao Pan, Taghrid Saad Alomar, Xiangyang Ji, Xiaojian Wang

    Published 2025-07-01
    “…A generative pretrained transformer model is subsequently developed for template-free retrosynthesis planning by pre-training on 10 billion generated data. …”
    Get full text
    Article
  2. 2

    ReactionT5: a pre-trained transformer model for accurate chemical reaction prediction with limited data by Tatsuya Sagawa, Ryosuke Kojima

    Published 2025-08-01
    “…This study introduces ReactionT5, a transformer-based chemical reaction foundation model pre-trained on the Open Reaction Database—a large publicly available reaction dataset. …”
    Get full text
    Article
  3. 3
  4. 4

    CGM: Copy Mechanism GPT with Mask for Ellipsis and Anaphora Resolution in Dialogue by Ji-Won Cho, Jinyoung Oh, Jeong-Won Cha

    Published 2024-12-01
    “…GPT (Generative Pre-trained Transformer) is a generative language model that demonstrates outstanding performance in the field of text generation. …”
    Get full text
    Article
  5. 5
  6. 6
  7. 7
  8. 8
  9. 9

    Vision Transformer untuk Klasifikasi Kematangan Pisang by Arya Pangestu, Bedy Purnama, Risnandar Risnandar

    Published 2024-02-01
    “…Penelitian dilakukan dengan menggunakan lima model ViT yang sudah dilatih sebelumnya atau pre-trained, yaitu ViT-B/16, ViT-B/32, ViT-L/16, ViT-L/32, and ViT-H/14 pada ImageNet-21k dan ImageNet-1k. …”
    Get full text
    Article
  10. 10

    Perbandingan Pretrained Model Transformer pada Deteksi Ulasan Palsu by Aisyah Awalina, Fitra Abdurrachman Bachtiar, Fitri Utaminingrum

    Published 2022-06-01
    “…Ada dua pendekatan yang dapat dilakukan dalam model Transformer yaitu pre-training dan fine-tuning. Penelitian sebelumnya telah banyak menggunakan fine-tuning dari model Transformer dikarenakan adanya kemudahan dalam pelatihan, waktu yang lebih sedikit, biaya dan kebutuhan lingkungan yang lebih rendah dibanding proses pre-training. …”
    Get full text
    Article
  11. 11

    CLFormer: a cross-lingual transformer framework for temporal forgery localization by Haonan Cheng, Hanyue Liu, Juanjuan Cai, Long Ye

    Published 2025-07-01
    “…Additionally, we introduced a cross-lingual transformer framework (CLFormer), which prioritizes audio features and utilizes a pre-trained multi-lingual Wav2Vec2 to enhance cross-lingual generalization, while incorporating visual features to further refine TFL. …”
    Get full text
    Article
  12. 12

    Lexicon-enhanced transformer with spatial-aware integration for Chinese named entity recognition by Jiachen Huang, Shuo Liu

    Published 2025-07-01
    “…Abstract Chinese Named Entity Recognition (CNER) is a fundamental and crucial task in information extraction. In recent years, pre-trained language and lexicon-based models have proven more powerful than the previous character-based models in CNER tasks. …”
    Get full text
    Article
  13. 13

    Visual language transformer framework for multimodal dance performance evaluation and progression monitoring by Lei Chen

    Published 2025-08-01
    “…We propose a novel transformer-based visual-language framework for multi-modal dance performance evaluation and progression monitoring. …”
    Get full text
    Article
  14. 14

    Herbify: an ensemble deep learning framework integrating convolutional neural networks and vision transformers for precise herb identification by Farhan Sheth, Ishika Chatter, Manvendra Jasra, Gireesh Kumar, Richa Sharma

    Published 2025-07-01
    “…Utilizing transfer learning, the research harnessed pre-trained Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs), then integrated these models into an ensemble framework that leverages the unique strengths of each architecture. …”
    Get full text
    Article
  15. 15
  16. 16

    FSID: a novel approach to human activity recognition using few-shot weight imprinting by Mohammad Belal, Taimur Hassan, Abdelfatah Hassan, Divya Velayudhan, Noureldin Elhendawi, Ahmad Aljarah, Irfan Hussain

    Published 2025-07-01
    “…This paper proposes Few-Shot Imprinted DINO (FSID), a novel framework for HAR in low-data regimes, combining Few-Shot learning with weight imprinting and a self-supervised vision transformer, DINO (Distillation with No Labels). The FSID pipeline begins by converting raw time-series sensor data (e.g., EMG and IMU signals) into spectrogram images using the Short-Time Fourier Transform. …”
    Get full text
    Article
  17. 17

    MATSFT: User query-based multilingual abstractive text summarization for low resource Indian languages by fine-tuning mT5 by Siginamsetty Phani, Ashu Abdul, M. Krishna Siva Prasad, V. Dinesh Reddy

    Published 2025-08-01
    “…Experimental results show that MATSFT outperforms the monolingual transformer model, pre-trained MTM, mT5 model, NLI model, IndicBART, mBART25, and mBART50 on the IL dataset. …”
    Get full text
    Article
  18. 18

    Rumor detection using dual embeddings and text-based graph convolutional network by Barsha Pattanaik, Sourav Mandal, Rudra M. Tripathy, Arif Ahmed Sekh

    Published 2024-11-01
    “…This model uses dual embedding from two pre-trained transformer models: generative pre-trained transformers (GPT) and bidirectional encoder representations from transformers (BERT). …”
    Get full text
    Article
  19. 19

    Unveiling the spectrum of Arabic offensive language: Taxonomy and insights. by Chaya Liebeskind, Yossef Haim Shrem, Marina Litvak, Natalia Vanetik

    Published 2025-01-01
    “…We have also analyzed the performance of pre-trained and fine-tuned Arabic transformer offensive language detection models on these datasets. …”
    Get full text
    Article
  20. 20

    High-accuracy physical property prediction for pure organics via molecular representation learning: bridging data to discovery by Qi Ou, Hongshuai Wang, Minyang Zhuang, Shangqian Chen, Lele Liu, Ning Wang, Zhifeng Gao

    Published 2025-07-01
    “…Traditional trial-and-error methods for discovering highly functional organic compounds are expensive and time-consuming. We employed a 3D transformer-based molecular representation learning algorithm to create the Org-Mol pre-trained model, using 60 million semi-empirically optimized small organic molecule structures. …”
    Get full text
    Article