Publications

Select a paper and use the Abstract or BibTeX buttons to reveal details.

  • Trust The Typical

    Debargha Ganguly, Sreehari Sankar, Biyao Zhang, Vikash Singh, Kanan Gupta, Harshini Kavuru, Alan Luo, Weicong Chen, Warren Morningstar, Raghu Machiraju, Vipin Chaudhary

    arXiv • Feb 2026

    Preprint
  • VERGE: Formal Refinement and Guidance Engine for Verifiable LLM Reasoning

    Vikash Singh, Darion Cassel, Nathaniel Weir, Nick Feng, Sam Bayless

    arXiv • Jan 2026

    Preprint
  • Mid-Think: Training-Free Intermediate-Budget Reasoning via Token-Level Triggers

    Yang Wang, Debargha Ganguly, Xinpeng Li, Chaoda Song, Shouren Wang, Vikash Singh, Vipin Chaudhary, Xiaotian Han

    arXiv • Jan 2026

    Preprint
  • Grammars of Formal Uncertainty: When to Trust LLMs in Automated Reasoning Tasks

    Ganguly, D., Singh, V., Sankar, S., Zhang, B., Zhang, X., Iyengar, S., Han, X., Sharma, A., Kalyanaraman, S., Chaudhary, V.

    NeurIPS 2025 • May 2025

    HTML arXiv Code
  • K⁴: Online Log Anomaly Detection Via Unsupervised Typicality Learning

    Singh, V.*, Chen, W.*, Rahmani, Z., Ganguly, D., Hariri, M., Chaudhary, V.

    *Equal contribution

    HiPC 2025 • July 2025

    HTML arXiv Code
  • Efficient Fine-Grained GPU Performance Modeling for Distributed Deep Learning of LLM

    Zhang, B., Zheng, M., Ganguly, D., Zhang, X., Singh, V., Chaudhary, V., Zhang, Z.

    HiPC 2025 • September 2025

    HTML arXiv

Research Interests

  • Formal Reasoning and Verification: Developing rigorous formal logic methodologies, leveraging SMT‐LIB encodings and solver frameworks to verify, interpret, and enhance the correctness of LLM-generated reasoning.
  • Fine-Tuning LLMs and Vision Models: Leveraging techniques such as LoRA to optimize large language and vision models for specific tasks while maintaining computational efficiency.
  • Redundancy Mitigation in LLMs: Investigating approaches to reduce redundancy in large language models, enhancing performance and efficiency.
  • Model Optimization: Developing strategies for optimizing machine learning models, including pruning and hyperparameter tuning, to improve both accuracy and resource utilization.
  • Explainable AI: Advancing interpretability in AI models, focusing on enhancing transparency and providing actionable insights for users.

Research Statement

Bridging Probabilistic Generative AI with Rigorous Formal Verification

My research centers on a fundamental tension in modern Artificial Intelligence: the gap between the powerful but hallucination-prone creativity of Large Language Models (LLMs) and the strict, deterministic guarantees required for trustworthy systems. I am developing a new class of "Neuro-Symbolic" architectures that do not just generate code or proofs, but actively reason about their own uncertainty, verify their outputs against formal constraints, and optimize their "thinking" budgets for maximum efficiency.

Current Contributions: Quantifying Uncertainty & Controlling Reasoning

My recent work, including research published at NeurIPS 2025, tackles the "epistemological gap" between probabilistic models and formal logic. I introduced a Probabilistic Context-Free Grammar (PCFG) framework to model the uncertainty of LLM-generated formal artifacts (like SMT-LIB programs). By treating LLM outputs not as final answers but as hypotheses with measurable uncertainty, I developed "selective verification" protocols that reduce logical errors by 14–100%.

Beyond verification, I focus on the efficiency of reasoning. In my work on "Mid-Think," I demonstrated that reasoning behaviors in hybrid models are driven by specific token-level triggers rather than high-level instructions. I leveraged this to create training-free prompting strategies that dynamically adjust the model's "compute budget" during inference, achieving superior accuracy-latency trade-offs. I also work on LLM safety using out-of-distribution (OOD) detection techniques.

Future Directions: Diffusion & Energy-Based Reasoning

I am currently pivoting towards Reasoning Diffusion Language Models and Energy-Based Models (EBMs) to overcome the limitations of standard auto-regressive generation. My hypothesis is that "reasoning" should not be a linear, left-to-right process, but an iterative refinement—similar to how diffusion models denoise an image.

Diffusion for Logic: I am exploring how diffusion processes can allow models to "revise" their logic in continuous latent space, enabling self-correction before generating a final answer.

Energy-Based Verification: I am investigating EBMs to model the "global consistency" of a reasoning chain. Instead of predicting the next token, these models assess the "energy" (or compatibility) of an entire proof or plan, guiding the generator toward formally correct states.

Impact & Vision

Drawing on my experience as an Applied Scientist Intern at AWS and my background in formal methods (Lean/Coq), my goal is to build AI systems that are safe enough for critical infrastructure. I aim to create models that don't just "guess" the answer, but construct a verifiable path to it—combining the flexibility of deep learning with the rigor of mathematical proof.