Scholar
Aakriti Agrawal
Google Scholar ID: 2RRnCRMAAAAJ
University of Maryland, College Park
LLM Superalignment
LLM alignment
Generative AI
Reinforcement Learning
Multi-Agent Systems
Follow
Homepage
↗
Google Scholar
↗
Citations & Impact
All-time
Citations
292
H-index
6
i10-index
6
Publications
14
Co-authors
18
list available
Contact
Email
agrawal5@umd.edu
CV
Open ↗
GitHub
Open ↗
LinkedIn
Open ↗
Publications
6 items
Towards Mitigating Hallucinations in Large Vision-Language Models by Refining Textual Embeddings
2025
Cited
0
Uncertainty-Aware Answer Selection for Improved Reasoning in Multi-LLM Systems
2025
Cited
0
EnsemW2S: Enhancing Weak-to-Strong Generalization with Large Language Model Ensembles
2025
Cited
0
PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models
2025
Cited
0
EnsemW2S: Can an Ensemble of LLMs be Leveraged to Obtain a Stronger LLM?
arXiv.org · 2024
Cited
7
Easy2Hard-Bench: Standardized Difficulty Labels for Profiling LLM Performance and Generalization
Neural Information Processing Systems · 2024
Cited
7
Resume (English only)
Academic Achievements
Paper 'Uncertainty-Aware Answer Selection for Improved Reasoning in Multi-LLM Systems' accepted at EMNLP 2025
'EnsemW2S: Can an Ensemble of SoTA LLMs be Leveraged to Obtain a Stronger LLM?' accepted at NeurIPS 2024 SafeGenAI workshop; under review at AAAI 2026
Co-authored 'Easy2Hard-Bench' published in NeurIPS 2024 Dataset Track
'Robustness to Multi-Modal Environment Uncertainty in MARL using Curriculum Learning' published at NeurIPS 2023 MASEC workshop
Co-authored 'WAVES: Benchmarking the Robustness of Image Watermarks' published at ICML 2024
Co-authored 'PoisonedParrot' accepted at NAACL 2025 and presented at SafeGenAI @ NeurIPS 2024
'Towards Mitigating Hallucinations in Large Vision-Language Models by Refining Textual Embeddings' under review
Background
4th-year Computer Science PhD student at the University of Maryland, College Park
Research interests include: improving reasoning in LLMs, reducing reward hacking in reasoning LLMs
Weak-to-strong generalization and self-improvement
LLM/VLM alignment, reasoning, and hallucination reduction
Reinforcement Learning, Multi-Agent Systems, and Uncertainty Estimation
Co-authors
18 total
Furong Huang
Associate Professor of Computer Science, University of Maryland
Mucong Ding
Department of Computer Science, University of Maryland
Bang An
University of Maryland, College Park
Tom Goldstein
Volpi-Cupal Professor of Computer Science, University of Maryland
Sicheng Zhu
Member of Technical Staff, OpenAI
Jordan K Terry
CEO, Farama Foundation
Co-author 7
Yuxin Wen
University of Maryland
×
Welcome back
Sign in to Agora
Welcome back! Please sign in to continue.
Email address
Password
Forgot password?
Continue
Do not have an account?
Sign up