Pranav Nair
Scholar

Pranav Nair

Google Scholar ID: Vr9vvooAAAAJ
Indian Institute of Technology (Banaras Hindu University), Varanasi
Natural Language Processing
Citations & Impact
All-time
Citations
715
 
H-index
5
 
i10-index
3
 
Publications
13
 
Co-authors
0
 
Resume (English only)
Academic Achievements
  • 1. Tandem Transformers for Inference Efficient LLMs (ICML 2024)
  • 2. CDQuant: Greedy Coordinate Descent for Accurate LLM Quantization (Arxiv 2024)
Research Experience
  • Pre-Doctoral Researcher at Google DeepMind, India, working with Dr. Praneeth Netrapalli, Dr. Arun Suggala, and Dr. Prateek Jain. His work involves making LLM inference faster through quantization, speculative decoding, and sparsification. He is also working on speeding up “million-context-attention” through clustering and approximate logit computation.
Background
  • Interests include but are not limited to: Making LLM inference faster through next-generation architectures, quantization, sparsification, speculative decoding, KV cache compression, adaptive routing for elastic models, etc., and speeding up LLM pretraining and finetuning through better adapters, novel loss functions, better second-order optimizers and faster checkpointing.
Co-authors
0 total
Co-authors: 0 (list not available)