Scholar
Satyapriya Krishna
Google Scholar ID: Q5bfPlkAAAAJ
Harvard University
Trustworthy AI
Large Language Models
Explainable & Fair ML
Follow
Homepage
↗
Google Scholar
↗
Citations & Impact
All-time
Citations
2,494
H-index
19
i10-index
25
Publications
20
Co-authors
0
Contact
Email
spkrishnaofficial@gmail.com
CV
Open ↗
Twitter
Open ↗
GitHub
Open ↗
LinkedIn
Open ↗
Publications
12 items
Evaluating Nova 2.0 Lite model under Amazon's Frontier Model Safety Framework
2026
Cited
2
From Narrow Unlearning to Emergent Misalignment: Causes, Consequences, and Containment in LLMs
2025
Cited
0
Self-Correcting Large Language Models: Generation vs. Multiple Choice
2025
Cited
0
The Alignment Auditor: A Bayesian Framework for Verifying and Refining LLM Objectives
2025
Cited
0
Learning from Failures: Understanding LLM Alignment through Failure-Aware Inverse RL
2025
Cited
0
D-REX: A Benchmark for Detecting Deceptive Reasoning in Large Language Models
2025
Cited
0
Evaluating the Critical Risks of Amazon's Nova Premier under the Frontier Model Safety Framework
2025
Cited
0
AILuminate: Introducing v1.0 of the AI Risk and Reliability Benchmark from MLCommons
arXiv.org · 2025
Cited
0
Load more
Resume (English only)
Background
AI Researcher with a focus on the trustworthy aspects of generative models, including explainability, fairness, privacy, and robustness.
Miscellany
No information provided about personal interests or hobbies
Co-authors
0 total
Co-authors: 0 (list not available)
×
Welcome back
Sign in to Agora
Welcome back! Please sign in to continue.
Email address
Password
Forgot password?
Continue
Do not have an account?
Sign up