Passionate about AI with roots in theoretical neuroscience,
I explore LLMs, AI agent, and human mobility—driven by a vision of accessible, everyday AI for all.
Engineering advanced AI systems—from autonomous multi-agent systems and scaling reasoning-focused LLMs on multi-node GPU clusters to performance profiling and distilling DeepSeek R1.
Building intelligent AI agents that dynamically reason, retrieve, and self-correct—from Agentic RAG with colocated vLLM inference to tool-augmented reasoning on the GAIA benchmark.
Key projects:
smolagents framework with Langfuse telemetry; achieves 80% accuracy on GAIA benchmark, outperforming GPT-4’s 14.4% baseline. Details →Explore all AI Agent projects →
Systematic performance analysis of Transformer architectures—benchmarking FP32 vs. BF16 mixed precision and profiling compute- vs. memory-bound operations in self-attention.
Key projects:
Explore all Benchmarking projects →
Advanced post-training and fine-tuning across large language models—from distilling DeepSeek R1 on multi-node HPC to instruction-tuning Llama 3.
Key projects:
Explore all LLM Distillation & Fine-Tuning projects →
Developing and fine-tuning generative models for image synthesis, as well as applying advanced deep learning architectures to real-world predictive modeling and audio processing tasks.
Key projects:
Explore all Generative AI & Applied Machine Learning projects →