Domain adaptation challenges of BERT in tokenization and sub-word representations of Out-of-Vocabulary words (original) (raw)
Related papers
Lessons Learned from Applying off-the-shelf BERT: There is no Silver Bullet
2020
ThisIsCompetition at SemEval-2019 Task 9: BERT is unstable for out-of-domain samples
Proceedings of the 13th International Workshop on Semantic Evaluation
LNLF-BERT: Transformer for Long Document Classification with Multiple Attention Levels
IEEE Access, 2024
Efficient Domain Adaptation of Language Models via Adaptive Tokenization
ArXiv, 2021
2021
GiBERT: Enhancing BERT with Linguistic Information using a Lightweight Gated Injection Method
Findings of the Association for Computational Linguistics: EMNLP 2021, 2021
BERT Probe: A python package for probing attention based robustness evaluation of BERT models
Software Impacts
ConvBERT: Improving BERT with Span-based Dynamic Convolution
2020
Domain Effect Investigation for Bert Models Fine-Tuned on Different Text Categorization Tasks
Arabian Journal for Science and Engineering, 2023
On the Prunability of Attention Heads in Multilingual BERT
ArXiv, 2021
Augmenting BERT Carefully with Underrepresented Linguistic Features
arXiv (Cornell University), 2020
TiltedBERT: Resource Adjustable Version of BERT
2022
Distilling Task-Specific Knowledge from BERT into Simple Neural Networks
HUBERT Untangles BERT to Improve Transfer across NLP Tasks
ArXiv, 2019
MemBERT: Injecting Unstructured Knowledge into BERT
ArXiv, 2021
How Can BERT Help Lexical Semantics Tasks?
arXiv (Cornell University), 2019
Noisy Text Data: Achilles’ Heel of BERT
Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020), 2020
What Does BERT Look at? An Analysis of BERT’s Attention
Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, 2019
A Review on BERT and Its Implementation in Various NLP Tasks
Advances in computer science research, 2023
Sequential Attention Module for Natural Language Processing
ArXiv, 2021
Improving the BERT model for long text sequences in question answering domain
Mareeswari Venkatachala, IJAAS Journal
International Journal of Advances in Applied Sciences (IJAAS), 2023
UoB at SemEval-2020 Task 12: Boosting BERT with Corpus Level Information
2020
How Far Does BERT Look At:Distance-based Clustering and Analysis of BERTś Attention
2020
On the Language-specificity of Multilingual BERT and the Impact of Fine-tuning
Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, 2021
A Primer in BERTology: What We Know About How BERT Works
Transactions of the Association for Computational Linguistics, 2020
Rethinking of BERT Sentence Embedding for Text Classification
Research Square (Research Square), 2024
LEA: Improving Sentence Similarity Robustness to Typos Using Lexical Attention Bias
Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
SenseBERT: Driving Some Sense into BERT
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Do Attention Heads in BERT Track Syntactic Dependencies?
ArXiv, 2019
Job Descriptions Keyword Extraction using Attention based Deep Learning Models with BERT
2021 3rd International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), 2021
Cornell University - arXiv, 2020
RoBERTa: A Robustly Optimized BERT Pretraining Approach
arXiv (Cornell University), 2019
KR-BERT: A Small-Scale Korean-Specific Language Model
2020
Diagnosing BERT with Retrieval Heuristics
Lecture Notes in Computer Science, 2020
Bangla-BERT: Transformer-Based Efficient Model for Transfer Learning and Language Understanding
IEEE Access