Word Order Matters When You Increase Masking (original) (raw)

Masked Language Modeling and the Distributional Hypothesis: Order Word Matters Pre-training for Little

Adina Williams

ArXiv, 2021

View PDFchevron_right

An Exploration of Placeholding in Neural Machine Translation

Orlando Chura

2019

View PDFchevron_right

COMPARATIVE ANALYSIS OF TRANSFORMER BASED LANGUAGE MODELS

Computer Science & Information Technology (CS & IT) Computer Science Conference Proceedings (CSCP)

View PDFchevron_right

Overview of the Transformer-based Models for NLP Tasks

anthony gillioz

Proceedings of the 2020 Federated Conference on Computer Science and Information Systems, 2020

View PDFchevron_right

On Robustness of Finetuned Transformer-based NLP Models

Venkateswara Rao Kagita

arXiv (Cornell University), 2023

View PDFchevron_right

A Comparison of Neural Models for Word Ordering

Eva Hasler

Proceedings of the 10th International Conference on Natural Language Generation

View PDFchevron_right

On Losses for Modern Language Models

Stéphane Aroca-Ouellette, Frank Rudzicz

Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020

View PDFchevron_right

End-to-End Transformer-Based Models in Textual-Based NLP

Abir Rahali

AI

View PDFchevron_right

Larger-Scale Transformers for Multilingual Masked Language Modeling

Naman goyal

Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021), 2021

View PDFchevron_right

BERTAC: Enhancing Transformer-based Language Models with Adversarially Pretrained Convolutional Neural Networks

Julien Kloetzer

Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)

View PDFchevron_right

AxFormer: Accuracy-driven Approximation of Transformers for Faster, Smaller and more Accurate NLP Models

Sanchari Sen

arXiv (Cornell University), 2020

View PDFchevron_right

Overview of the progression of state-of-the-art language models

TELKOMNIKA JOURNAL

TELKOMNIKA Telecommunication Computing Electronics and Control, 2024

View PDFchevron_right

Out of Order: How important is the sequential order of words in a sentence in Natural Language Understanding tasks?

Thắng Phạm

Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2021

View PDFchevron_right

Language Modeling with Deep Transformers

Albert Zeyer

Interspeech 2019

View PDFchevron_right

Investigating semantic subspaces of Transformer sentence embeddings through linear structural probing

Dmitry Nikolaev

BlackboxNLP, 2023

View PDFchevron_right

Analyzing Encoded Concepts in Transformer Language Models

Firoj Alam

Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies

View PDFchevron_right

Representation biases in sentence transformers

Dmitry Nikolaev

Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, 2023

View PDFchevron_right

Neural Language Models are not Born Equal to Fit Brain Data, but Training Helps

Christophe Pallier

HAL (Le Centre pour la Communication Scientifique Directe), 2022

View PDFchevron_right

Masked Language Model Scoring

Julian Salazar

Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics

View PDFchevron_right

BERT4SO: Neural Sentence Ordering by Fine-tuning BERT

shengchao liu

2021

View PDFchevron_right

Which Sentence Embeddings and Which Layers Encode Syntactic Structure?

Jesús Calvillo Tinoco

2020

View PDFchevron_right

Serial Recall Effects in Neural Language Modeling

Hassan Hajipoor

Proceedings of the 2019 Conference of the North

View PDFchevron_right

Empirical Evaluation of Pre-trained Transformers for Human-Level NLP: The Role of Sample Size and Dimensionality

Huy Vu

Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021

View PDFchevron_right

Analysis of sentence embedding models using prediction tasks in natural language processing

Yossi Adi

IBM Journal of Research and Development, 2017

View PDFchevron_right

Stress Test Evaluation of Transformer-based Models in Natural Language Understanding Tasks

Andres Carvallo

arXiv (Cornell University), 2020

View PDFchevron_right

Exploring Neural Language Models via Analysis of Local and Global Self-Attention Spaces

Senja Pollak

2021

View PDFchevron_right

Improving N-gram Language Models with Pre-trained Deep Transformer

Yutong Pang

2019

View PDFchevron_right

Pre-training Polish Transformer-Based Language Models at Scale

Rafał Poświata

Artificial Intelligence and Soft Computing, 2020

View PDFchevron_right