AI, for Less
I strongly advocate for "AI, for Less". My research shows you do not always need massive computational power — simpler, optimized models can yield exceptional results while remaining transparent and trustworthy. We emphasize reproducibility and simplicity over "black box" deep learning, and show that lightweight active learning, co-training, and warm-starts can match or beat expensive alternatives.
- FSE'26 Data-Light SE Challenge [pdf]
- VERIFAI'26 Exploiting Software Sparsity [pdf]
- TOSEM'26 LLMs Warming-Up Active Learning [pdf]
- CACM'25 The Case for Compact AI [pdf]
- TOSEM'24 Learning from Very Little Data
- EMSE'24 Co-Training for Defect Prediction
- IEEE Access'24 Partial Ordering for Model Reasoning
- EMSE'22 Minimizing Tech Debt Labeling Cost
LLMs & Deep Learning
Critical assessments of where LLMs and deep learning help — and where they hurt — in software engineering tasks. We study how to prompt, warm-start, and robustify LLM-based SE pipelines, and identify when simpler baselines still win.
- EMSE'26 Improving LLM Annotations [pdf]
- MSR'26 Domain Knowledge for LLM Optimization [pdf]
- IEEE Internet Comput.'23 Variances in Robust DL
- TSE'22 Oversampling for DL Defect Prediction
- MSR'22 Improving DL for SE Analytics
- MSR'22 GANs for Security Class Imbalance
Security & Vulnerability
Practical methods for vulnerability detection and cyberthreat intelligence, using active learning and adversarial ML to help teams prioritize security efforts without drowning in false positives.
- KAIS'25 Mining CTI Attack Patterns [pdf]
- ICDM'24 Temporal Cyberattack Patterns
- EMSE'22 Ensembles Against Evasion Attacks
- EMSE'21 Security Bug Report Classification
Defect Prediction & Software Analytics
Building better predictors for software quality, project health, and technical debt — using smarter sampling, transfer learning, and hyperparameter optimization rather than brute-force compute. Includes benchmark repositories for reproducible research across hundreds of SE tasks.
- MSR'26 Multi-Objective Optimization Repo [pdf]
- TSE'25 Is HPO Different for SE? [pdf]
- TSE'25 Static Code Mining Retrospective [pdf]
- SIGSOFT SEN'25 Replications & Negative Results
- ESA'23 Expert System for Cloud SE
- EMSE'22 Process vs. Product Metrics
- EMSE'22 Predicting OSS Project Health
- MSR'22 Stabilizing Models Across Projects
- ICSE'21 Early Life Cycle Defect Prediction
Accountability & Trust
AI software must meet the same correctness and auditability standards as any safety-critical system. We develop methods to verify, audit, and correct AI outputs — ensuring decisions can be inspected, reproduced, and challenged — and probe the fragility of causal reasoning in high-stakes environments.
- JSS'26 Explaining Optimization Heuristics [pdf]
- IEEE SW'25 AI in State Courts [pdf]
- EMSE'25 Causal Graphs in SE [pdf]
- EMSE'23 Interpreting Model-Based Opt.
- TSE'24 Equalized Odds via Pre-Processing
- IEEE Access'24 Scalability, Privacy, & Performance