Plex: Towards Reliability using Pretrained Large Model Extensions (original) (raw)
Related papers
Beyond Sole Strength: Customized Ensembles for Generalized Vision-Language Models
arXiv (Cornell University), 2023
Enabling Calibration In The Zero-Shot Inference of Large Vision-Language Models
arXiv (Cornell University), 2023
Task Residual for Tuning Vision-Language Models
arXiv (Cornell University), 2022
ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models
arXiv (Cornell University), 2022
A Survey of Vision-Language Pre-Trained Models
Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence
To what extent do DNN-based image classification models make unreliable inferences?
Empirical Software Engineering, 2021
Uncertainty in AI: Evaluating Deep Neural Networks on Out-of-Distribution Images
Uncertainty in AI: Evaluating Deep Neural Networks on Out-of-Distribution Images, 2023
Analyzing the Domain Robustness of Pretrained Language Models, Layer by Layer
2021
A Comprehensive Evaluation Framework for Deep Model Robustness
2021
ConfidenceBench: A Confidence Calibration Benchmark for Large Language Models
ImageNet-X: Understanding Model Mistakes with Factor of Variation Annotations
arXiv (Cornell University), 2022
Discrepancies among pre-trained deep neural networks: a new threat to model zoo reliability
Proceedings of the 30th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering
ECO: Ensembling Context Optimization for Vision-Language Models
arXiv (Cornell University), 2023
Robust Learning Through Cross-Task Consistency
2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020
Retrieval Augmentation to Improve Robustness and Interpretability of Deep Neural Networks
2021
REBUS: A Robust Evaluation Benchmark of Understanding Symbols
arXiv (Cornell University), 2024
Making Pre-trained Language Models End-to-end Few-shot Learners with Contrastive Prompt Tuning
Cornell University - arXiv, 2022
GluonCV and GluonNLP: Deep Learning in Computer Vision and Natural Language Processing
J. Mach. Learn. Res., 2020
LAION-5B: An open large-scale dataset for training next generation image-text models
Cornell University - arXiv, 2022
Developing reliability metrics and validation tools for datasets with deep linguistic information
2011
Assessing the Robustness of Visual Question Answering
2019
Proceedings of the 15th International Conference on Knowledge Technologies and Data-driven Business
2021
Examining Large Pre-Trained Language Models for Machine Translation: What You Don't Know About It
2022
Toward Improving the Robustness of Deep Learning Models via Model Transformation
Proceedings of the 37th IEEE/ACM International Conference on Automated Software Engineering
Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters
arXiv (Cornell University), 2024
WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training
ArXiv, 2021
Data-centric Reliability Evaluation of Individual Predictions
Cornell University - arXiv, 2022
Semantically Distributed Robust Optimization for Vision-and-Language Inference
2021
arXiv (Cornell University), 2024
Bag of Experts Architectures for Model Reuse in Conversational Language Understanding
Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 3 (Industry Papers)
cViL: Cross-Lingual Training of Vision-Language Models using Knowledge Distillation
Cornell University - arXiv, 2022
Exploring Uncertainty Measures for Image-caption Embedding-and-retrieval Task
ACM Transactions on Multimedia Computing, Communications, and Applications, 2021
Measuring and Mitigating Local Instability in Deep Neural Networks
arXiv (Cornell University), 2023