Syntactic Structure Distillation Pretraining for Bidirectional Encoders
Saved in:
Main Authors: | Adhiguna Kuncoro, Lingpeng Kong, Daniel Fried, Dani Yogatama, Laura Rimell, Chris Dyer, Phil Blunsom |
---|---|
Format: | Article |
Language: | English |
Published: |
The MIT Press
2024-08-01
|
Series: | Transactions of the Association for Computational Linguistics |
Online Access: | http://dx.doi.org/10.1162/tacl_a_00345 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
by: Laurent Sartran, et al.
Published: (2022-12-01) -
Relational Memory-Augmented Language Models
by: Qi Liu, et al.
Published: (2022-05-01) -
Heterogeneous federated bidirectional knowledge distillation transfer semi-supervised modulation recognition
by: Peihan QI, et al.
Published: (2023-11-01) -
<b>Learning Machine Translation Cyril Goutte†, Nicola Cancedda*, Marc Dymetman*, and George Foster† (editors)</b> (†Institute for Information Technology, National Research Council Canada; *Xerox Research Centre Europe) Cambridge, MA: The MIT Press, 2009, xii+316 pp; hardbound, ISBN 978-0-262-07297-7, $45.00, £29.95
by: Phil Blunsom
Published: (2021-11-01) -
Pretraining Enhanced RNN Transducer
by: Junyu Lu, et al.
Published: (2024-12-01)