MIRRORWIC: On Eliciting Word-in-Context Representations from Pretrained Language Models
View / Open Files
Authors
Liu, Q
Liu, F
Collier, N
Korhonen, A
Vulić, I
Publication Date
2021-09-19Journal Title
CoNLL 2021 - 25th Conference on Computational Natural Language Learning, Proceedings
Conference Name
25th Conference on Computational Natural Language Learning (CoNLL 2021)
ISBN
9781955917056
Publisher
Association for Computational Linguistics
Type
Conference Object
This Version
VoR
Metadata
Show full item recordCitation
Liu, Q., Liu, F., Collier, N., Korhonen, A., & Vulić, I. (2021). MIRRORWIC: On Eliciting Word-in-Context Representations from Pretrained Language Models. CoNLL 2021 - 25th Conference on Computational Natural Language Learning, Proceedings https://doi.org/10.17863/CAM.78495
Abstract
Recent work indicated that pretrained language models (PLMs) such as BERT and
RoBERTa can be transformed into effective sentence and word encoders even via
simple self-supervised techniques. Inspired by this line of work, in this paper
we propose a fully unsupervised approach to improving word-in-context (WiC)
representations in PLMs, achieved via a simple and efficient WiC-targeted
fine-tuning procedure: MirrorWiC. The proposed method leverages only raw texts
sampled from Wikipedia, assuming no sense-annotated data, and learns
context-aware word representations within a standard contrastive learning
setup. We experiment with a series of standard and comprehensive WiC benchmarks
across multiple languages. Our proposed fully unsupervised MirrorWiC models
obtain substantial gains over off-the-shelf PLMs across all monolingual,
multilingual and cross-lingual setups. Moreover, on some standard WiC
benchmarks, MirrorWiC is even on-par with supervised models fine-tuned with
in-task data and sense labels.
Keywords
cs.CL, cs.CL
Identifiers
External DOI: https://doi.org/10.17863/CAM.78495
This record's URL: https://www.repository.cam.ac.uk/handle/1810/331050
Statistics
Total file downloads (since January 2020). For more information on metrics see the
IRUS guide.
Recommended or similar items
The current recommendation prototype on the Apollo Repository will be turned off on 03 February 2023. Although the pilot has been fruitful for both parties, the service provider IKVA is focusing on horizon scanning products and so the recommender service can no longer be supported. We recognise the importance of recommender services in supporting research discovery and are evaluating offerings from other service providers. If you would like to offer feedback on this decision please contact us on: support@repository.cam.ac.uk