-
1
RSGPT: a generative transformer model for retrosynthesis planning pre-trained on ten billion datapoints
Published 2025-07-01“…A generative pretrained transformer model is subsequently developed for template-free retrosynthesis planning by pre-training on 10 billion generated data. …”
Get full text
Article -
2
ReactionT5: a pre-trained transformer model for accurate chemical reaction prediction with limited data
Published 2025-08-01“…This study introduces ReactionT5, a transformer-based chemical reaction foundation model pre-trained on the Open Reaction Database—a large publicly available reaction dataset. …”
Get full text
Article -
3
Unsupervised evaluation of pre-trained DNA language model embeddings
Published 2025-08-01Get full text
Article -
4
CGM: Copy Mechanism GPT with Mask for Ellipsis and Anaphora Resolution in Dialogue
Published 2024-12-01“…GPT (Generative Pre-trained Transformer) is a generative language model that demonstrates outstanding performance in the field of text generation. …”
Get full text
Article -
5
Octascope: A Lightweight Pre-Trained Model for Optical Coherence Tomography
Published 2025-01-01“…This multi-domain pre-training enables Octascope to generalize across varied tissue types. …”
Get full text
Article -
6
Generative AI-powered multilingual ASR for seamless language-mixing transcriptions
Published 2025-07-01Subjects: Get full text
Article -
7
Ensemble Approach for Image Recompression-Based Forgery Detection
Published 2024-01-01Get full text
Article -
8
Impact of retrieval augmented generation and large language model complexity on undergraduate exams created and taken by AI agents
Published 2025-01-01Subjects: Get full text
Article -
9
Vision Transformer untuk Klasifikasi Kematangan Pisang
Published 2024-02-01“…Penelitian dilakukan dengan menggunakan lima model ViT yang sudah dilatih sebelumnya atau pre-trained, yaitu ViT-B/16, ViT-B/32, ViT-L/16, ViT-L/32, and ViT-H/14 pada ImageNet-21k dan ImageNet-1k. …”
Get full text
Article -
10
Perbandingan Pretrained Model Transformer pada Deteksi Ulasan Palsu
Published 2022-06-01“…Ada dua pendekatan yang dapat dilakukan dalam model Transformer yaitu pre-training dan fine-tuning. Penelitian sebelumnya telah banyak menggunakan fine-tuning dari model Transformer dikarenakan adanya kemudahan dalam pelatihan, waktu yang lebih sedikit, biaya dan kebutuhan lingkungan yang lebih rendah dibanding proses pre-training. …”
Get full text
Article -
11
CLFormer: a cross-lingual transformer framework for temporal forgery localization
Published 2025-07-01“…Additionally, we introduced a cross-lingual transformer framework (CLFormer), which prioritizes audio features and utilizes a pre-trained multi-lingual Wav2Vec2 to enhance cross-lingual generalization, while incorporating visual features to further refine TFL. …”
Get full text
Article -
12
Lexicon-enhanced transformer with spatial-aware integration for Chinese named entity recognition
Published 2025-07-01“…Abstract Chinese Named Entity Recognition (CNER) is a fundamental and crucial task in information extraction. In recent years, pre-trained language and lexicon-based models have proven more powerful than the previous character-based models in CNER tasks. …”
Get full text
Article -
13
Visual language transformer framework for multimodal dance performance evaluation and progression monitoring
Published 2025-08-01“…We propose a novel transformer-based visual-language framework for multi-modal dance performance evaluation and progression monitoring. …”
Get full text
Article -
14
Herbify: an ensemble deep learning framework integrating convolutional neural networks and vision transformers for precise herb identification
Published 2025-07-01“…Utilizing transfer learning, the research harnessed pre-trained Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs), then integrated these models into an ensemble framework that leverages the unique strengths of each architecture. …”
Get full text
Article -
15
-
16
FSID: a novel approach to human activity recognition using few-shot weight imprinting
Published 2025-07-01“…This paper proposes Few-Shot Imprinted DINO (FSID), a novel framework for HAR in low-data regimes, combining Few-Shot learning with weight imprinting and a self-supervised vision transformer, DINO (Distillation with No Labels). The FSID pipeline begins by converting raw time-series sensor data (e.g., EMG and IMU signals) into spectrogram images using the Short-Time Fourier Transform. …”
Get full text
Article -
17
MATSFT: User query-based multilingual abstractive text summarization for low resource Indian languages by fine-tuning mT5
Published 2025-08-01“…Experimental results show that MATSFT outperforms the monolingual transformer model, pre-trained MTM, mT5 model, NLI model, IndicBART, mBART25, and mBART50 on the IL dataset. …”
Get full text
Article -
18
Rumor detection using dual embeddings and text-based graph convolutional network
Published 2024-11-01“…This model uses dual embedding from two pre-trained transformer models: generative pre-trained transformers (GPT) and bidirectional encoder representations from transformers (BERT). …”
Get full text
Article -
19
Unveiling the spectrum of Arabic offensive language: Taxonomy and insights.
Published 2025-01-01“…We have also analyzed the performance of pre-trained and fine-tuned Arabic transformer offensive language detection models on these datasets. …”
Get full text
Article -
20
High-accuracy physical property prediction for pure organics via molecular representation learning: bridging data to discovery
Published 2025-07-01“…Traditional trial-and-error methods for discovering highly functional organic compounds are expensive and time-consuming. We employed a 3D transformer-based molecular representation learning algorithm to create the Org-Mol pre-trained model, using 60 million semi-empirically optimized small organic molecule structures. …”
Get full text
Article