Ashish Hooda

Portrait


I am a Research Scientist at Google DeepMind in Gemini team. I work on post-training with as focus on improving Gemini's instruction following capabilities.

I completed my PhD in 2025 at University of Wisconsin-Madison majoring in Computer Sciences. Before that, I received my undergraduate from Indian Institute of Technology Delhi.

Email: ashish1995hooda@gmail.com.

Recent News

Work Experience


Google DeepMind: Research Scientist
May 2025 - Present
Gemini GenAI, Post-training, Instruction Following
Working on improving Gemini’s instruction following capabilities.

Google : Research Intern
July 2023 - Nov 2023
Host : Mihai Christodorescu and Miltos Allamanis
Internship with the Android Security and Learning for Code teams. Worked on evaluating program semantics understanding of Large Language Models for Code.
Amazon AWS : Applied Scientist Intern
June 2022 - Sept 2022
Hosts : Ali Torkamani
Internship with the AWS Security Analytics and AI Research team. Worked on efficient training of Graph Neural Network for intrusion detection on billion node scale graphs.

Research


What Really is a Member? Discrediting Membership Inference via Poisoning
NeurIPS 2025 (The Thirty-Ninth Annual Conference on Neural Information Processing Systems)
PAPER
Neal Mangaokar*, Ashish Hooda*, Zhuohang Li, Bradley A Malin, Kassem Fawaz, Somesh Jha, Atul Prakash, Amrita Roy Chowdhury


PRP
PAPER
Zi Wang*, Divyam Anshumaan*, Ashish Hooda, Yudong Chen, Somesh Jha



PAPER
Andrey Labunets, Nishit Pandya, Ashish Hooda, Xiaohan Fu, Earlence Fernandes


PRP
PAPER
Ashish Hooda, Rishabh Khandelwal, Prasad Chalasani, Kassem Fawaz, Somesh Jha


PRP
Guruprasad V Ramesh, Harrison Rosenberg, Ashish Hooda, Kassem Fawaz


PRP
Neal Mangaokar*, Ashish Hooda*, Jihye Choi, Shreyas Chandrashekaran, Kassem Fawaz, Somesh Jha, Atul Prakash


CodeLLMs
Ashish Hooda, Mihai Christodorescu, Miltos Allamanis, Aaron Wilson, Kassem Fawaz, Somesh Jha.


Phash Surveillance
Experimental Analyses of the Physical Surveillance Risks in Client-Side Content Scanning
NDSS 2024 (Network and Distributed System Security Symposium)
Ashish Hooda, Andrey Labunets, Tadayoshi Kohno, Earlence Fernandes.


Detecting Diffusion Deepfakes
D4: Detection of Adversarial Diffusion Deepfakes Using Disjoint Ensembles
WACV 2024 (IEEE/CVF Winter Conference on Applications of Computer Vision)


OARS
Stateful Defenses for Machine Learning Models Are Not Yet Secure Against Black- box Attacks
CCS 2023 (ACM Conference on Computer and Communications Security)


Stateful Theory


SkillFence
SkillFence: A Systems Approach to Mitigating Voice-Based Confusion Attacks
IMWUT / UBICOMP 2022 (ACM Interactive, Mobile, Wearable and Ubiquitous Technologies)
Ashish Hooda, Matthew Wallace, Kushal Jhunjhunwalla, Earlence Fernandes , Kassem Fawaz.


Invisible Perturbations
Invisible Perturbations: Physical Adv Examples Exploiting the Rolling Shutter Effect
CVPR 2021 (Conference on Computer Vision and Pattern Recognition)
Athena Sayles*, Ashish Hooda*, Mohit Gupta, Rahul Chatterjee, Earlence Fernandes.


Talks

Do Large Code Models Understand Programming Concepts? Counterfactual Analysis for Code Predicates
JetBrains Research, Oct 2024
Is Attack Detection A Viable Defense For Adversarial Machine Learning?
Visa Research, Jun 2024
Do Code LLMs understand program semantics?
Google Learning for Code Team, Nov 2023
Do Stateful Defenses Work Against Black-Box Attacks?
Google AI Red Team, Oct 2023
Deepfake Detection Against Adaptive Attackers
Google AI Red Team, Aug 2023