Nazneen Fatema Rajani


News

  • Invited talk at UC Berkeley Decentralization Summit [ slides ]
  • Keynote at the AI Conference [link]
  • Tutorial on Trustworthy Generative AI at ICML '23 and FAccT '23 [slides]
  • Fireside chat with Pear VC on Generative AI [blog]
  • Guest lectures at UT Austin, UMich Ann Arbor, and Georgia Tech March-April '23 [slides].
  • Invited talk at NIST on March 2, '23 [recording].
  • I'll be giving the EMNLP'22 Industry track keynote.
  • 5 papers accepted to the main track and 2 to the demo track of EMNLP '22.
  • Invited panelist at the NIST AI Risk management workshop.
  • Invited talk at EMNLP '22 EvoNLP workshop.

I am Research Lead at Hugging Face 🤗. Currently, I am part of the team working on building an open-source alternative to ChatGPT called H4. While building such a powerful LLM, I am thinking deeply about evaluation, red-teaming, alternatives to RLHF, and model interpretability in the era of LLMs. Check out my blogpost for some musings as well as empirical results on these topics.

My interests and expertise lie at the intersection of Model Evaluation, Robustness, and Interpretability. My long term research agenda includes 1. Dataset curation for understanding training and evaluation data, 2. Human-in-the-loop training and evaluation for model robustness, 3. Repurposing LLMs and PLMs to make them more usable in the real world. Checkout my HF spaces to see what I am building

Before joing Hugging Face, I was a Senior Research Scientist at Salesforce Research where I worked with Richard Socher and Caiming Xiong on commonsense reasoning and interpretability in NLP. I led a small team focused on building robust natural language generation models. Prior to working at Salesforce, I completed my Ph.D. thesis in the Department of Computer Science at the Machine Learning Research Group of the University of Texas, Austin. I worked with my advisor Prof. Ray Mooney on problems in NLP, Vision and at the intersection of NLP and Vision. I have also worked on problems in Explainable AI (XAI) wherein I proposed a scalable approach to generate visual explanations for ensemble methods using the localization maps of the component systems. Evaluating explanations is also a challenging problem and I proposed two novel evaluation metrics that does not require human generated GT.

I completed my MS in CS with thesis advised by Jason Baldridge on new topic detection using topical alignment from tweets based on their author and recipient.