I am a final-year PhD candidate at Virginia Tech, Blacksburg, advised by Dr. Daphne Yao, working at the intersection of security and generative AI. My research focuses on applying generative AI to improve real-world cybersecurity outcomes, spanning LLM safety, synthetic data generation, and adversarial ML. I build evaluation-driven systems that improve security robustness in practice and have published in IEEE S&P, Asia CCS, and ACSAC. My work includes hands-on experience with LLMs, GANs, and diffusion models for security-specific data generation and for mitigating toxicity in chatbot customization pipelines.
I am currently seeking industry research or postdoctoral roles starting in Summer 2026 and welcome inquiries via email.
Problem: Security ML classifiers are often limited by data scarcity and imbalance, rather than algorithmic design.
Approach: Applied Generative AI-based synthetic data augmentation to improve classifier generalization. Evaluated multiple state-of-the-art GenAI methods and introduced Nimai, a controlled synthetic data generation scheme for security tasks.
Evaluation: Conducted empirical evaluation across 7 diverse security tasks, including severely low-data settings.
Outcome: Achieved up to 32.6% performance improvement in security classifiers.
Problem: Fine-tuning LLMs on untrusted conversational data can introduce toxic or unsafe behavior.
Approach: Developed TuneShield, a defense framework for safe fine-tuning of LLaMA-based chat models, combining LLM-based toxicity detection, targeted synthetic data generation ("healing data"), and alignment using PEFT-based Direct Preference Optimization (DPO).
Evaluation: Evaluated against toxicity injection, adaptive adversarial, and jailbreak attacks during conversational fine-tuning.
Outcome: Reduced toxic outputs to near-zero levels while preserving conversational quality.
Problem: Manual vulnerability testing for third-party libraries does not scale for software supply-chain security.
Approach: Built a self-correcting multi-agent LLM system for automated JUnit security test generation and execution.
Evaluation: Evaluated across 50 Java client codebases, using LLM-based validation to determine whether generated tests successfully trigger known vulnerabilities.
Outcome: Achieved automated end-to-end vulnerability test generation and validation using self-correcting LLM agents.
Graduate Research Assistant
Applied Research Collaboration
AI Security & Generative AI
AI Security, LLM Safety & Alignment, Generative AI, Adversarial Machine Learning, Security Evaluation & Benchmarking
Model Fine-Tuning & Alignment
Direct Preference Optimization (DPO), PEFT (LoRA), Supervised Fine-Tuning (SFT), Instruction Tuning, Prompt Engineering
Synthetic Data & Robustness
Synthetic Data Generation, Tabular Data Augmentation, Concept Drift, Robustness Analysis, Data Poisoning & Toxicity Injection
Agent-Based Systems
LLM Agents, Multi-Agent Systems, Self-Correcting Agents, Automated Security Test Generation, Agentic Reasoning
Security Domains
Malware & Phishing Detection, Malicious URL Detection, Network Intrusion Detection (IDS), Software Supply-Chain Security, Access Control Systems
ML Frameworks & Tooling
PyTorch, Hugging Face Transformers, LangChain, Evaluation Pipelines