Word Order Matters When You Increase Masking (original) (raw)
Related papers
ArXiv, 2021
An Exploration of Placeholding in Neural Machine Translation
2019
COMPARATIVE ANALYSIS OF TRANSFORMER BASED LANGUAGE MODELS
Computer Science & Information Technology (CS & IT) Computer Science Conference Proceedings (CSCP)
Overview of the Transformer-based Models for NLP Tasks
Proceedings of the 2020 Federated Conference on Computer Science and Information Systems, 2020
On Robustness of Finetuned Transformer-based NLP Models
arXiv (Cornell University), 2023
A Comparison of Neural Models for Word Ordering
Proceedings of the 10th International Conference on Natural Language Generation
On Losses for Modern Language Models
Stéphane Aroca-Ouellette, Frank Rudzicz
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020
End-to-End Transformer-Based Models in Textual-Based NLP
AI
Larger-Scale Transformers for Multilingual Masked Language Modeling
Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021), 2021
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
arXiv (Cornell University), 2020
Overview of the progression of state-of-the-art language models
TELKOMNIKA Telecommunication Computing Electronics and Control, 2024
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2021
Language Modeling with Deep Transformers
Interspeech 2019
BlackboxNLP, 2023
Analyzing Encoded Concepts in Transformer Language Models
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Representation biases in sentence transformers
Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, 2023
Neural Language Models are not Born Equal to Fit Brain Data, but Training Helps
HAL (Le Centre pour la Communication Scientifique Directe), 2022
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
BERT4SO: Neural Sentence Ordering by Fine-tuning BERT
2021
Which Sentence Embeddings and Which Layers Encode Syntactic Structure?
2020
Serial Recall Effects in Neural Language Modeling
Proceedings of the 2019 Conference of the North
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021
Analysis of sentence embedding models using prediction tasks in natural language processing
IBM Journal of Research and Development, 2017
Stress Test Evaluation of Transformer-based Models in Natural Language Understanding Tasks
arXiv (Cornell University), 2020
Exploring Neural Language Models via Analysis of Local and Global Self-Attention Spaces
2021
Improving N-gram Language Models with Pre-trained Deep Transformer
2019
Pre-training Polish Transformer-Based Language Models at Scale
Artificial Intelligence and Soft Computing, 2020