I recently earned my PhD in Computer Science from Virginia Tech, Blacksburg, where I was advised by Dr. Daphne Yao. I work at the intersection of security and generative AI, turning advanced AI methods into practical defenses for real-world cybersecurity problems. My research spans LLM safety, synthetic data generation, and adversarial ML, with publications in IEEE S&P, Asia CCS, and ACSAC. I build evaluation-driven systems that improve security robustness in practice, drawing on hands-on experience with LLMs, GANs, and diffusion models for security-specific data generation and safer chatbot customization.
I am currently seeking industry research or postdoctoral roles starting in Summer 2026 and welcome inquiries via email.
Problem: Security ML classifiers are often limited by data scarcity and imbalance, rather than algorithmic design.
Approach: Applied Generative AI-based synthetic data augmentation to improve classifier generalization. Evaluated multiple state-of-the-art GenAI methods and introduced Nimai, a controlled synthetic data generation scheme for security tasks.
Evaluation: Conducted empirical evaluation across 7 diverse security tasks, including severely low-data settings.
Outcome: Achieved up to 32.6% performance improvement in security classifiers.
Problem: Fine-tuning LLMs on untrusted conversational data can introduce toxic or unsafe behavior.
Approach: Developed TuneShield, a defense framework for safe fine-tuning of LLaMA-based chat models, combining LLM-based toxicity detection, targeted synthetic data generation ("healing data"), and alignment using PEFT-based Direct Preference Optimization (DPO).
Evaluation: Evaluated against toxicity injection, adaptive adversarial, and jailbreak attacks during conversational fine-tuning.
Outcome: Reduced toxic outputs to near-zero levels while preserving conversational quality.
Problem: Manual vulnerability testing for third-party libraries does not scale for software supply-chain security.
Approach: Built a self-correcting multi-agent LLM system for automated JUnit security test generation and execution.
Evaluation: Evaluated across 50 Java client codebases, using LLM-based validation to determine whether generated tests successfully trigger known vulnerabilities.
Outcome: Achieved automated end-to-end vulnerability test generation and validation using self-correcting LLM agents.
Graduate Research Assistant
Applied Research Collaboration
AI Security & Generative AI
AI Security, LLM Safety & Alignment, Generative AI, Adversarial Machine Learning, Security Evaluation & Benchmarking
Model Fine-Tuning & Alignment
Direct Preference Optimization (DPO), PEFT (LoRA), Supervised Fine-Tuning (SFT), Instruction Tuning, Prompt Engineering
Synthetic Data & Robustness
Synthetic Data Generation, Tabular Data Augmentation, Concept Drift, Robustness Analysis, Data Poisoning & Toxicity Injection
Agent-Based Systems
LLM Agents, Multi-Agent Systems, Self-Correcting Agents, Automated Security Test Generation, Agentic Reasoning
Security Domains
Malware & Phishing Detection, Malicious URL Detection, Network Intrusion Detection (IDS), Software Supply-Chain Security, Access Control Systems
ML Frameworks & Tooling
PyTorch, Hugging Face Transformers, LangChain, Evaluation Pipelines