AI Engineer and researcher focused on mechanistic interpretability, LLM reliability, and high-performance ML systems.
I build and ship production AI systems, and I study how those systems work internally. My primary focus is mechanistic interpretability: tracing transformer circuits, testing causal hypotheses, and designing targeted interventions that improve behavior without losing efficiency.
I previously completed an M.S. in Artificial Intelligence at Rochester Institute of Technology (May 2025), where my work spanned multimodal deepfake detection, LLM evaluation and robustness, and performance-aware training workflows.
Activation tracing, attention-head profiling, and intervention-based analysis to identify task-relevant transformer mechanisms and measure causal contribution to final model behavior.
Benchmark-driven evaluation across GPT-4, Claude, Llama, and Mistral families with stress testing, retrieval-augmented prompting, and regression checks for reasoning and factual consistency.
Latency/throughput optimization, profiling, and regression debugging for production AI services, including model serving, asynchronous orchestration, and memory-aware optimization.
Multimodal CNN-transformer architectures for lip-sync manipulation detection, synthetic data generation, and robust out-of-distribution evaluation under realistic deployment constraints.