Publications
Select a paper and use the Abstract or BibTeX buttons to reveal details.
-
Trust The Typical
arXiv • Feb 2026
-
VERGE: Formal Refinement and Guidance Engine for Verifiable LLM Reasoning
arXiv • Jan 2026
-
Mid-Think: Training-Free Intermediate-Budget Reasoning via Token-Level Triggers
arXiv • Jan 2026
-
Grammars of Formal Uncertainty: When to Trust LLMs in Automated Reasoning Tasks
NeurIPS 2025 • May 2025
-
K⁴: Online Log Anomaly Detection Via Unsupervised Typicality Learning
*Equal contribution
HiPC 2025 • July 2025
-
Efficient Fine-Grained GPU Performance Modeling for Distributed Deep Learning of LLM
HiPC 2025 • September 2025
Research Interests
- Formal Reasoning and Verification: Developing rigorous formal logic methodologies, leveraging SMT‐LIB encodings and solver frameworks to verify, interpret, and enhance the correctness of LLM-generated reasoning.
- Fine-Tuning LLMs and Vision Models: Leveraging techniques such as LoRA to optimize large language and vision models for specific tasks while maintaining computational efficiency.
- Redundancy Mitigation in LLMs: Investigating approaches to reduce redundancy in large language models, enhancing performance and efficiency.
- Model Optimization: Developing strategies for optimizing machine learning models, including pruning and hyperparameter tuning, to improve both accuracy and resource utilization.
- Explainable AI: Advancing interpretability in AI models, focusing on enhancing transparency and providing actionable insights for users.
Research Statement
My research centers on a fundamental tension in modern Artificial Intelligence: the gap between the powerful but hallucination-prone creativity of Large Language Models (LLMs) and the strict, deterministic guarantees required for trustworthy systems. I am developing a new class of "Neuro-Symbolic" architectures that do not just generate code or proofs, but actively reason about their own uncertainty, verify their outputs against formal constraints, and optimize their "thinking" budgets for maximum efficiency.
My recent work, including research published at NeurIPS 2025, tackles the "epistemological gap" between probabilistic models and formal logic. I introduced a Probabilistic Context-Free Grammar (PCFG) framework to model the uncertainty of LLM-generated formal artifacts (like SMT-LIB programs). By treating LLM outputs not as final answers but as hypotheses with measurable uncertainty, I developed "selective verification" protocols that reduce logical errors by 14–100%.
Beyond verification, I focus on the efficiency of reasoning. In my work on "Mid-Think," I demonstrated that reasoning behaviors in hybrid models are driven by specific token-level triggers rather than high-level instructions. I leveraged this to create training-free prompting strategies that dynamically adjust the model's "compute budget" during inference, achieving superior accuracy-latency trade-offs. I also work on LLM safety using out-of-distribution (OOD) detection techniques.
I am currently pivoting towards Reasoning Diffusion Language Models and Energy-Based Models (EBMs) to overcome the limitations of standard auto-regressive generation. My hypothesis is that "reasoning" should not be a linear, left-to-right process, but an iterative refinement—similar to how diffusion models denoise an image.
Diffusion for Logic: I am exploring how diffusion processes can allow models to "revise" their logic in continuous latent space, enabling self-correction before generating a final answer.
Energy-Based Verification: I am investigating EBMs to model the "global consistency" of a reasoning chain. Instead of predicting the next token, these models assess the "energy" (or compatibility) of an entire proof or plan, guiding the generator toward formally correct states.
Drawing on my experience as an Applied Scientist Intern at AWS and my background in formal methods (Lean/Coq), my goal is to build AI systems that are safe enough for critical infrastructure. I aim to create models that don't just "guess" the answer, but construct a verifiable path to it—combining the flexibility of deep learning with the rigor of mathematical proof.