Thanks for stopping by — I'm really glad you're here.
I grew up in India, fascinated by technology and what it could do. A few years back, I made the leap to move to the United States to study Computer Engineering — and that decision changed everything. Today, I'm a Senior Machine Learning Engineer building production ML systems that serve millions of people in healthcare.
My work spans the full ML lifecycle: feature engineering, model training, MLOps pipelines, HIPAA compliance, and production monitoring on AWS. Outside of healthcare, I build things that genuinely fascinate me — like QuantEdge v6.0, a live institutional quant analytics platform combining 8 ML models.
But there's more to my story than just work.
I believe life gets interesting when you stay curious about everything.
When I'm not working, you'll find me:
Each of these passions fuels my creativity and makes me better at what I do.
Thanks for visiting. Feel free to explore and stay in touch.
Six years building and deploying ML systems in enterprise healthcare has taught me that the hardest part isn't the model — it's making it production-grade, trustworthy, and something a team can actually ship on. I specialize in the full journey from research to deployment: fine-tuning LLMs with Hugging Face, building agentic workflows with LangChain, and training production models in PyTorch. Whether it's an evaluation framework, a secure inference pipeline, or a multi-agent system, I care about building AI that enterprise teams can rely on in the real world.
Institutional quantitative AI platform combining 8 ML models: LSTM temporal forecasting, XGBoost/LightGBM ensemble, 5-state HMM regime detection, GJR-GARCH volatility, FinBERT NLP sentiment, and 100K-path Monte Carlo simulation. Deployed live on AWS ECS Fargate with Redis, PostgreSQL, CloudFront CDN, and Terraform IaC.
🚀 Live Demo 🔗 GitHubHybrid retrieval-augmented generation system combining knowledge graph reasoning with semantic vector search using Amazon Bedrock (Claude 3). Designed weighted graph + vector ranking to improve contextual accuracy and explainability. Implemented evaluation metrics (Precision@K, Recall@K, MRR) and stage-wise latency instrumentation for performance visibility.
🔗 GitHubProduction-style real-time ML feature store and low-latency inference system. Kafka-compatible streaming ingestion (Redpanda) with dual-store architecture (Redis for online serving, Parquet for offline training). Multi-worker FastAPI inference service with end-to-end training and deployment pipeline.
🔗 GitHubCloud-native platform architected for streaming-first, agent-driven data ingestion and enrichment. Autonomous agent logic enriches and routes events before persistence — decisions before storage. Designed for both analytical queries and GenAI workloads, addressing real-time decisioning in regulated, cost-sensitive environments where latency, governance, and transactional consistency all matter simultaneously.
🔗 GitHubEnterprise AI security middleware — LangChain + FastAPI gateway with prompt injection detection, PII scrubbing, and rate limiting. Multi-provider LLM routing (OpenAI, Anthropic, Azure OpenAI) with RBAC, token-budget enforcement, and HIPAA-compliant audit logging for enterprise multi-tenant deployments.
🔗 GitHubProduction healthcare ML systems break not from algorithmic failures but because something upstream changed quietly — a vendor pushed a schema migration, the annual ICD-10-CM update redistributed code frequencies, a new health center came onto the feed. Standard monitoring tells you something is wrong. This work tells you which source, and why, in 23 minutes instead of 4 hours.
The core insight is mathematical: KL divergence decomposes additively under product measures. Any monitor watching the aggregate feature vector discards the per-source terms that make attribution possible. Moving monitoring to the source-table boundary — before any join — recovers it. This paper formalizes that principle, proves when attribution succeeds and when simultaneous multi-source shifts make it provably impossible (Corollary 1), and builds a complete four-component system — ASDM — that runs end-to-end in 130ms.
Why it matters beyond healthcare: Any ML system assembling features from heterogeneous independent sources — financial services, public-sector analytics, digital advertising — faces the same attribution problem. The taxonomy mechanisms are domain-specific. The monitoring principle — move to the source boundary before the join — is not.
Capturing moments. Exploring perspectives.
Gallery coming soon. Stay tuned for aerial views, travel stories, and moments worth capturing.
Beyond ML engineering and photography, I'm exploring entrepreneurship with a close friend. The intersection of AI, data, and real-world problems is where the most interesting opportunities live.
More details coming soon.