Paper 'DeepPolar: Inventing Nonlinear Large-Kernel Polar Codes via Deep Learning' accepted to ICML 2024; Paper 'Task-Aware Distributed Source Coding Under Dynamic Bandwidth' published in NeurIPS 2023; Paper 'Nested Construction of Polar Codes via Transformers' accepted to ISIT 2024; Paper 'TinyTurbo: Efficient Turbo Decoders on Edge' accepted to ISIT 2022; Awarded the Agnes T. and Charles F. Wiebusch Fellowship from the Cockrell School of Engineering for 2025-2026.
Research Experience
Worked as a Research Engineer in the Physical Layer Modelling team at Qualcomm for two years; Currently pursuing Ph.D. at the University of Texas at Austin, collaborating with Pramod Viswanath (Princeton), Krishna Narayanan (TAMU), Sewoong Oh (UW Seattle), Sandeep Chinchali (UT Austin).
Education
Ph.D. candidate in the ECE department at the University of Texas at Austin, advised by Hyeji Kim; Graduated with Bachelors and Masters in Electrical Engineering from IIT Madras in 2019, Master’s Thesis advisors were Dr. Andrew Thangaraj and Dr. Radha Krishna Ganti.
Background
Broadly interested in problems at the intersection of Deep Learning and Information Theory, focusing on using vision and language foundation model priors to build ultra-low rate image compression schemes. Also interested in deep learning for tabular datasets and generative models for synthetic datasets.
Miscellany
In spare time, enjoys reading books, watching documentaries, and hiking.