Domain adaptation challenges of BERT in tokenization and sub-word representations of Out-of-Vocabulary words (original) (raw)

Lessons Learned from Applying off-the-shelf BERT: There is no Silver Bullet

Victor Makarenkov

2020

View PDFchevron_right

ThisIsCompetition at SemEval-2019 Task 9: BERT is unstable for out-of-domain samples

Changki Lee

Proceedings of the 13th International Workshop on Semantic Evaluation

View PDFchevron_right

LNLF-BERT: Transformer for Long Document Classification with Multiple Attention Levels

Linh Manh Pham

IEEE Access, 2024

View PDFchevron_right

Efficient Domain Adaptation of Language Models via Adaptive Tokenization

Jason Kessler

ArXiv, 2021

View PDFchevron_right

The heads hypothesis: A unifying statistical approach towards understanding multi-headed attention in BERT

Madhura Pande

2021

View PDFchevron_right

GiBERT: Enhancing BERT with Linguistic Information using a Lightweight Gated Injection Method

Maria Liakata

Findings of the Association for Computational Linguistics: EMNLP 2021, 2021

View PDFchevron_right

BERT Probe: A python package for probing attention based robustness evaluation of BERT models

Mahnoor Shahid

Software Impacts

View PDFchevron_right

ConvBERT: Improving BERT with Span-based Dynamic Convolution

Zihang Jiang

2020

View PDFchevron_right

Domain Effect Investigation for Bert Models Fine-Tuned on Different Text Categorization Tasks

Ferhat Bozkurt

Arabian Journal for Science and Engineering, 2023

View PDFchevron_right

On the Prunability of Attention Heads in Multilingual BERT

Madhura Pande

ArXiv, 2021

View PDFchevron_right

Augmenting BERT Carefully with Underrepresented Linguistic Features

Jekaterina Novikova

arXiv (Cornell University), 2020

View PDFchevron_right

TiltedBERT: Resource Adjustable Version of BERT

Mohammad Sharifkhani

2022

View PDFchevron_right

Distilling Task-Specific Knowledge from BERT into Simple Neural Networks

Melison Dylan

View PDFchevron_right

HUBERT Untangles BERT to Improve Transfer across NLP Tasks

Paul Smolensky

ArXiv, 2019

View PDFchevron_right

MemBERT: Injecting Unstructured Knowledge into BERT

Federico Ruggeri

ArXiv, 2021

View PDFchevron_right

How Can BERT Help Lexical Semantics Tasks?

leyang cui

arXiv (Cornell University), 2019

View PDFchevron_right

Noisy Text Data: Achilles’ Heel of BERT

Piyush Makhija

Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020), 2020

View PDFchevron_right

What Does BERT Look at? An Analysis of BERT’s Attention

Christopher D Manning

Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, 2019

View PDFchevron_right

A Review on BERT and Its Implementation in Various NLP Tasks

ANKITA THOMBRE

Advances in computer science research, 2023

View PDFchevron_right

Sequential Attention Module for Natural Language Processing

lianxin jiang

ArXiv, 2021

View PDFchevron_right

Improving the BERT model for long text sequences in question answering domain

Mareeswari Venkatachala, IJAAS Journal

International Journal of Advances in Applied Sciences (IJAAS), 2023

View PDFchevron_right

UoB at SemEval-2020 Task 12: Boosting BERT with Corpus Level Information

Harish Tayyar Madabushi

2020

View PDFchevron_right

How Far Does BERT Look At:Distance-based Clustering and Analysis of BERTś Attention

Jingwen Leng

2020

View PDFchevron_right

On the Language-specificity of Multilingual BERT and the Impact of Fine-tuning

Lonneke van der Plas

Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, 2021

View PDFchevron_right

A Primer in BERTology: What We Know About How BERT Works

Ольга Ковалева

Transactions of the Association for Computational Linguistics, 2020

View PDFchevron_right

Rethinking of BERT Sentence Embedding for Text Classification

mona farouk

Research Square (Research Square), 2024

View PDFchevron_right

LEA: Improving Sentence Similarity Robustness to Typos Using Lexical Attention Bias

Mario Almagro

Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining

View PDFchevron_right

SenseBERT: Driving Some Sense into BERT

Or Dagan

Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics

View PDFchevron_right

Do Attention Heads in BERT Track Syntactic Dependencies?

Shikha Bordia

ArXiv, 2019

View PDFchevron_right

Job Descriptions Keyword Extraction using Attention based Deep Learning Models with BERT

Hussain Mahdi

2021 3rd International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), 2021

View PDFchevron_right

BERT's output layer recognizes all hidden layers? Some Intriguing Phenomena and a simple way to boost BERT

Wei-tsung Kao

Cornell University - arXiv, 2020

View PDFchevron_right

RoBERTa: A Robustly Optimized BERT Pretraining Approach

Naman Goyal

arXiv (Cornell University), 2019

View PDFchevron_right

KR-BERT: A Small-Scale Korean-Specific Language Model

Suzi Park

2020

View PDFchevron_right

Diagnosing BERT with Retrieval Heuristics

Arthur Câmara

Lecture Notes in Computer Science, 2020

View PDFchevron_right

Bangla-BERT: Transformer-Based Efficient Model for Transfer Learning and Language Understanding

Abdullah As Sami 1704111

IEEE Access

View PDFchevron_right