Hi, I'm Aman!
I'm an AI Researcher at Foundation-AI (Cisco, part of Robust Intelligence startup acquisition) specializing in foundation models for security applications, particularly reasoning, long-horizon planning, and agentic systems. I've trained 8B-20B parameter models across multi-node GPU clusters (50+ GPUs). My research spans AI for security, AI security, privacy-preserving ML, and LLM safety, with work deployed in production models that have seen 300K+ downloads (in 2025 alone). My AI Safety work has been featured in SC Magazine, The Register, and other outlets, and led to invitations to OpenAI's Red Teaming Network and Anthropic's Model Safety Bug Bounty Program.
With a Masters in Privacy Engineering from Carnegie Mellon University, I've published at venues like USENIX and AAAI. I worked with Professor Norman Sadeh on LLM security research and Niloofar Mireshghallah on privacy-preserving ML. Currently, I build specialized RL environments for security domains, including custom CTF environments for automated penetration testing curriculum training and vulnerability detection frameworks for iterative code patch discovery.
scrollable: please scroll left or right →
Cisco AI researcher Aman Priyanshu is developing autonomous vulnerability search agents using SLMs that mimic human analyst workflows through iterative investigation...
Cisco BlogsCisco Foundation AI's first open-source security base model, trained by Aman Priyanshu & team.
SC MediaMeta's Prompt-Guard, is vulnerable to a simple exploit with a 99.8% success rate... AI Security Researcher Aman Priyanshu wrote in a blog post...
Cisco BlogsRelease of the instruction-tuned Foundation-sec-8B, developed by Aman Priyanshu & team, enabling ready-to-use security assistance and analysis.
Communications of the ACMPriyanshu said the biggest risk is organizations assuming their jailbreaking defenses are 100% effective.
Cisco BlogsThe first open-weight security reasoning model, trained by Aman Priyanshu & team using custom RLVR environments.
Cisco & TILOS Faculty TalksExploring reasoning models, their training, common pitfalls, and solutions as an AI Researcher @ Cisco.
VKTR...adopt privacy-preserving technologies such as differential privacy or fine-tuning with synthetic data... as Priyanshu explained.
The Trace...breaking a single prohibited request into several innocuous-seeming questions can make an AI system more than five times as likely to produce a harmful response - Supriti & Aman
Cisco BlogsENUM-based attack achieved an ASR of 52.89%, compared to 12.44% for normal API calling... - Aman
Equinox ITAman Priyanshu and Supriti Vijay analysed expert activations in gpt-oss-20b and pruned under-utilised experts across domain-specialised variants... With this work I imagine pruning will get more attention...
The Koffee Conversation Show...got into privacy preserving ML optimization @ Eder Labs and AI Security @ Robust Intelligence...
Supriti Vijay*, Aman Priyanshu*, Anu Vellore, Baturay Saglam, Amin Karbasi
arXiv Preprint / Accepted at Conference on Applied Machine Learning in Information Security (CAMLIS)
Paul Kassianik, Baturay Saglam, Alexander Chen, Blaine Nelson, Anu Vellore, Massimo Aufiero, Fraser Burch, Dhruv Kedia, Avi Zohary, Sajana Weerawardhena, Aman Priyanshu, Adam Swanda, Amy Chang, Hyrum Anderson, Kojin Oshiba, Omar Santos, Yaron Singer, Amin Karbasi
Technical Report (Foundation-AI, Cisco)
Aman Priyanshu, Yash Maurya, Vy Tran, Suriya Ganesh Ayyamperumal
USENIX Conference on Privacy Engineering Practice and Respect
Aman Priyanshu*, Supriti Vijay*
arXiv Preprint
Aman Priyanshu, Yash Maurya, Zuofei Hong
arXiv Preprint
Aman Priyanshu*, Supriti Vijay*, Ayush Kumar, Rakshit Naidu, Fatemehsadat Mireshghallah
arXiv Preprint
Aman Priyanshu*, Supriti Vijay*
arXiv Preprint
Supriti Vijay*, Aman Priyanshu*
Updatable Machine Learning Workshop, ICML 2022
Aman Priyanshu, Rakshit Naidu, Fatemehsadat Mireshghallah, Mohammad Malekzadeh
Privacy Preserving Machine Learning Workshop, ACM CCS 2021
Rakshit Naidu*, Aman Priyanshu*, Aadith Kumar, Sasikanth Kotti, Haofan Wang, Fatemehsadat Mireshghallah
Responsible Computer Vision Workshop, CVPR 2021 & Privacy Preserving Machine Learning Workshop, ACM CCS 2021
scrollable: please scroll left or right →
Released 232 domain-specialized models from 4.2B to 20B (downloaded 30K+ times within 6 months) by analyzing expert activations and pruning.
We cracked a real-world differentially private synthetic data by linking public information to exposed PII overnight.
We built a keyword extractor, forgot about it, and somehow researchers are actually using it in their work.
I created a hierarchical topic modeling dataset from RedPajama, with 100k samples and 3 levels of topics.
I created a hierarchical topic modeling dataset from RedPajama, with 100k samples and 3 levels of topics.
I built a vanilla JavaScript library that lets you use AI APIs directly in browsers, no backend needed.
I mapped linguistic patterns in YC startup pitches like growing trees, and built a semantic search tool to explore them.
We broke AI safeguards by splitting harmful prompts into innocent sub-questions.
Feb 2023
Key Courses: Prompt Engineering (17730), AI Governance (17716), Deep Learning (11785), Computer Technology Law (17562), Differential Privacy (17731), Information Security (17631), & Usability (17734)
Key Courses: Data Structures and Algorithms, Design and Analysis of Algorithms, Object Oriented Programming, Probability and Statistics, Computer Networks, Operating Systems, Database Management
April 2025 - January 2026
Released and maintained production security LLMs (200K+ downloads) spanning Base, Instruct, and Reasoning 8B variants over a year-long development cycle. Created custom RLVR (Reinforcement Learning from Verifiable Rewards) cybersecurity environments for training reasoning capabilities. Continuously retrained and improved models based on user feedback, achieving state-of-the-art performance on CTI benchmarks while remaining deployable on-premise for sensitive security workflows. Trained 8B-20B parameter models across multi-node GPU clusters (50+ GPUs) using FSDP, DeepSpeed, and custom rig for MoEs.
August 2025
Built first-ever RL environment for iterative vulnerability retrieval with 4,500 GitHub Security Advisories. ARGUS-LLaMA-1B (43.9% success) outperforms GPT-4o (39.3%) and LLaMA-405B (41.8%) at 400× smaller size. Presented at CAMLIS 2025. Framework enables automated "reason, search, retrieve, repeat" workflows for discovering vulnerable code in large repositories.
August 2025
Released 232 domain-specialized models 4.2B to 20B (30K+ downloads within 6 months of release) by analyzing expert activation patterns in GPT-OSS-20B's Mixture-of-Experts architecture. Strategic pruning of underutilized experts maintains performance while reducing computational overhead. Includes interactive analytics dashboard and layer comparison tools for behavioral pattern analysis.
November 2025
Research on multi-turn retrieval architecture enabling compact models (350M-1.2B) to outperform systems 400× larger on general information retrieval domains like Science, Economics, Math, Programming / Coding, Robotics, General Knowledge, and BioMed. Combines synthetic trajectory generation, turn-level reinforcement learning, and beam search for adaptive information retrieval. Achieves 77.6% on SciFact (vs 72.6% prior SOTA) and 25.2% on BRIGHT (vs 22.1%). Demonstrates domain-agnostic generalizability of our synthetic search techniques.
October 2022
Built a Python library integrating semi-supervised attention for few-shot and zero-shot domain adaptation in keyphrase extraction. Library has been adopted by researchers across multiple domains for keyword extraction tasks.
August 2021
Tools for hyperparameter optimization in DP-SGD training. Proposed novel customizable reward function enabling users to define privacy-utility tradeoffs through single objective optimization.
August 2021
Unsupervised ML system for extracting contextually similar texts, applied to indexing academic literature, law precedents, and financial records. Won Code Innovation Series hackathon in association with GitHub.
July 2021
Real-time face verification system with obstruction detection, spoof detection, blur detection, and environment approval. Utilized deep neural networks and genetic algorithms for low-latency performance. Won 1st place in HackRx 2.0 by Bajaj Finserv.
2020 - 2025
15+ hackathon wins across ML/AI competitions including 1st place at Strong Compute (ARC AGI), HackRx (Bajaj Finserv), Code Innovation Series, ACM Datathon, CalHacks, etc.; runners up in IEEE BigMM, BobHacks, ShowYourSkill, etc. Projects spanned reasoning frameworks, speech processing, optimization, and privacy-preserving ML.
February 2023
Selected as one of twelve individuals for the AAAI UC program, recognizing my research on Privacy and Fairness.
June 2022
Selected as MITACS Globalink Research Scholar at Concordia University, Montreal, Canada for research collaboration in AI and Privacy-Preserving Machine Learning.
January 2020
Selected as one of the recipients of the Intel Edge AI Scholarship Program. Learned about Machine Learning Implementation on the Edge.
scrollable: please scroll left or right →
Understand and play with federated learning hyperparams! In-browser tensorflow-js simulation of FedAvg to understand and gain intuition about IID and Non-IID Federated Learning settings.
A unique twist on classic Tetris where players manage a privacy budget to reveal blocks, demonstrating differential privacy concepts through gameplay. Experience privacy-utility tradeoffs in an engaging way.
An interactive game exploring machine learning unlearning and fairness concepts. Players select data points that least impact the dataset, providing hands-on experience with data removal and model fairness considerations.