Allegro Lab (original) (raw)

logo/logo-vertical-hires.png

This page is under development.

The AI, Language, Learning, Generalization, and Robustness (ALLeGRo) Lab studies natural language processing and machine learning with a focus building reliable NLP systems for a wide range of scenarios. We aim for a deeper understanding of how NLP systems work, when they fail, and how they can be improved.

Here are the research questions we have been working on recently:

news

Sep 03, 2024 Welcome to the new Allegro Lab website.

selected publications

  1. AIES
    Operationalizing content moderation "accuracy" in the Digital Services Act
    2024
  2. ACL Findings
    Proving membership in LLM pretraining data via data watermarks
    Johnny Tian-Zheng Wei*, Ryan Yixiang Wang*, and Robin Jia
    2024
  3. NAACL
    Do Localization Methods Actually Localize Memorized Data in LLMs?
    Ting-Yun Chang, Jesse Thomason, and Robin Jia
  4. EMNLP
    Chain-of-Questions Training with Latent Answers for Robust Multistep Question Answering
    Wang Zhu, Jesse Thomason, and Robin Jia
  5. EMNLP Findings
    Estimating Large Language Model Capabilities without Labeled Test Data
    Harvey Yiyun Fu, Qinyuan Ye, Albert Xu, Xiang Ren, and Robin Jia
  6. EACL Findings
    Benchmarking Long-tail Generalization with Likelihood Splits
    Ameya Godbole Jia, and Robin
  7. EMNLP Findings
    Generalization Differences between End-to-End and Neuro-Symbolic Vision-Language Reasoning Systems
    Wang Zhu, Jesse Thomason, and Robin Jia
  8. ACL
    Selective Question Answering under Domain Shift
    Amita Kamath, Robin Jia, and Percy Liang
  9. NAACL
    Document-Level N-ary Relation Extraction with Multiscale Representation Learning
    Robin Jia, Cliff Wong, and Hoifung Poon
  10. EMNLP
    Adversarial Examples for Evaluating Reading Comprehension Systems
    Robin Jia, and Percy Liang
  11. EMNLP
    When Parts are Greater Than Sums: Individual LLM Components Can Outperform Full Models
    Ting-Yun Chang, Jesse Thomason, and Robin Jia
    2024
  12. NeurIPS
    Pre-trained Large Language Models Use Fourier Features to Compute Addition
    Tianyi Zhou, Deqing Fu, Vatsal Sharan, and Robin Jia
    2024
  13. arxiv
    Language Models can Infer Action Semantics for Classical Planners from Environment Feedback
    Wang Zhu, Ishika Singh, Robin Jia, and Jesse Thomason
    2024
  14. NeurIPS
    Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Models
    Deqing Fu, Tian-Qi Chen, Robin Jia, and Vatsal Sharan
    2024