I build AI systems that are actually in production - and I extract reusable engines from them so the next team doesn't rebuild what I already shipped.
Six years shipping data and AI systems across enterprise HR, food-tech, and fintech. Currently building the retrieval, workflow orchestration, and governed SQL layers that make AI useful in production - not just impressive in demos.
ExponentHR: Data platform engineering at scale - CI/CD ownership compressing deployment cycles 3 months to 14 days, CDC ETL -67% compute cost, payroll-critical AAG database automation on Azure.
Modular engines: extracted the tailoring core into tailor-resume (MCP server, PyPI, CLI, Streamlit) and the discovery layer into JobScout - platform thinking, not tool accumulation.
Six years building data and AI systems across enterprise, food-tech, and fintech.
Jul 2024 - Present
Data Engineer
ExponentHR · Addison, TX
Data platform engineering / CI/CD ownership
Compressed deployment cycles from 3 months to 14 days by owning CI/CD end-to-end through Azure DevOps, eliminating 11 weeks of cross-team idle time per release.
Reengineered CDC ETL from full-table reloads to incremental merge upserts: runtime 30min→8min, compute cost -67%.
Engineered one-click idempotent Azure DevOps pipeline for Contained AAG databases (restore, security, CDC, listener validation), eliminating ~1 hour per request across 20+ daily copy-downs.
AzureAzure DevOpsData PlatformCDCPythonSQL
Aug 2023 - Jul 2024
Data Engineer
Missouri University of Science and Technology · Rolla, MO
ML infrastructure / anomaly detection · M.S. Data Science - GPA 4.0 - Dec 2023
Engineered Azure AI Anomaly Detector pipelines selecting optimal algorithms per time-series profile, achieving 95%+ detection accuracy - caught a production memory leak 4 hours before outage.
Implemented tunable alert thresholds, filtering ~250 weekly non-actionable P3 alerts; signal-to-noise improved from 1:5 to 1:1.2.
Migrated from static D-Series VMs to AKS with HPA: CPU utilization 12%→64%, nodes consolidated 20→4-8 dynamic, Azure spend cut $3,200/month.
Azure AIAKS/HPAKubernetesAnomaly DetectionPython
Jun 2023 - Aug 2023
Engineering Intern
C2FO · Leawood, KS
Analyzed B2B financial transaction patterns via SQL to identify user behavior trends and inform product prioritization across the preferred offers tool.
Authored data-driven PRDs that reduced development resource allocation time by 50% by eliminating spec ambiguity across engineering, design, and business stakeholders.
SQLData AnalysisPRD
Sep 2020 - Mar 2021
Business Intelligence Analyst | Supply
udaan.com · India
Built predictive demand forecasting models driving $4 million annual savings and 7% ROI improvement through optimized inventory allocation.
Achieved 99.3% fulfillment rate through statistical capacity planning managing end-to-end first-mile to last-mile operations.
Developed automated ETL pipelines for financial modeling dashboards in Power BI.
ForecastingPower BISQLETLSupply Chain
Mar 2018 - Sep 2020
Business Analyst
Zomato · Hyderabad, India
Built real-time competitor analytics platform tracking pricing, delivery times, and coverage - fed pricing changes that contributed to 9% market share gain in contested metros.
Optimized search relevance with ranking models using contextual signals (time, location, cuisine affinity), improving search-to-conversion across millions of daily queries.
Built Elasticsearch enterprise search indexing 100K+ documents, reducing Support Desk volume ~80% through self-serve query resolution.
Missouri University of Science and Technology · Rolla, MO
GPA 4.0 / 4.0Jan 2022 - Dec 2023Taylor & Francis Publication
Certifications
DP-700 Fabric Data Engineer AssociateMicrosoft
Azure AI Fundamentals AI-900Microsoft
Generative AI FundamentalsDatabricks
Data Warehouse in Microsoft FabricApplied Skills · Microsoft
SQL AdvancedHackerRank
Certified Scrum Product OwnerScrum Alliance
Stack
Stack I work in daily.
Senior+ AI/ML platform tools, organized by domain. Hover or focus a tile for one-line context. Every tile is a tool I've shipped with — no aspirational filler.
AI, LLM & RAG
Anthropic SDKPython SDK for Claude API + tool use + prompt caching
ClaudePrimary LLM in production cascade (Sonnet 4.6 default)
GPT-4oOpenAI fallback in the multi-LLM cascade
Hugging FaceTransformers + sentence-embedding models for RAG
LangChainRAG orchestration where prompt-chain depth justifies it
MCPModel Context Protocol — shipped tailor-resume MCP server
Languages & Frameworks
PythonPrimary backend + ML language across every system
TypeScriptChrome MV3 extension + type-safe API contracts
RedisSession cache + queue, Upstash serverless in prod
KafkaStreaming ingest in Portfolio Risk Analytics
Lakehouse & Compute
SparkStructured Streaming consumer for Kafka topics
DatabricksLakehouse platform — production ML + analytics workloads
SSISSQL Server Integration Services — ETL pipelines
Delta LakeVersioned, ACID-safe lakehouse tables on Spark
IcebergOpen table format — petabyte-scale data lakes, time travel
Cloud & Deployment
AWSS3 + IAM + CloudWatch on production deployments
AzureAzure Functions + Storage in prior enterprise work
DockerMulti-stage builds for every shipped service
KubernetesHave used — staging clusters for AutoApply experiments
Fly.ioAutoApply AI production region — IAD + DFW
Observability & Ops
PrometheusMetrics scraping in containerized services
GrafanaDashboards for ingest lag + provider latency
SentryBackend + extension error tracking on AutoApply
MLflowExperiment tracking for fraud-detection pipelines
AirflowDAG orchestration in fraud-detection ML pipeline
GitHub ActionsCI on every repo — lint, types, tests, deploy
Key Accomplishments
Deployment velocity
Cycles cut from 3 months to 14 days by owning CI/CD end-to-end through Azure DevOps. Eliminated 11 weeks of cross-team idle time per release.
CDC ETL optimization
Reengineered CDC ETL from full-table reloads to incremental merge upserts: runtime 30 min to under 8 min, compute cost -67%.
Always-on platform
AutoApply AI runs live on Fly.io - zero cold start, 9-page React dashboard, 40+ FastAPI endpoints in production.
Four distribution surfaces
tailor-resume ships as CLI, Streamlit app, MCP server, and PyPI package - from one shared engine.
Technical Skills
Data platform
LLM-integrated data platform on Azure with governance enforced at the generation layer. RLS/CLS access control baked into the data product, not bolted on after.
Workflow stack
Chrome MV3, FastAPI, PostgreSQL, and Redis orchestrate the full discover → tailor → apply → track lifecycle.
ML infrastructure
Azure AI anomaly detection - 95%+ accuracy, AKS/HPA migration cutting Azure spend $3,200/month. Signal-to-noise from 1:5 to 1:1.2.
MCP protocol
tailor-resume is an MCP server - any Claude Code session or MCP-aware agent calls it directly, zero integration work. 97M monthly SDK downloads.
Projects & Open Source
Four production systems, each independently deployable.
Built end-to-end, tested in production, and modular enough that other teams can use the engines without adopting the whole stack.
FastAPI endpoints — verified by counting @router decorators in the AutoApply backend.
11
ATS adapters — eleven content-script adapters (Greenhouse, Lever, Workday, Ashby, SmartRecruiters, iCIMS, BambooHR, Jobvite, Taleo, Recruitee, Workable) live in the extension's ats/ directory.
6
LLM providers — a cascading provider chain across Claude, GPT-4o, Kimi, Ollama, Gemini, and Groq, all defined in the backend's providers service.
190
automated tests — one hundred ninety pytest cases in the AutoApply backend, verifiable with pytest collect-only.
Applying for jobs is a ritual that eats weeks without improving outcomes. I rebuilt the process from scratch - Chrome MV3 extension detects the job form, tailors the resume using an LLM cascade across 6 providers, submits through 11 different ATS systems, and logs everything to a React dashboard. No manual copy-paste. No lost applications. Live on Fly.io.
Figure: AutoApply AI topology. Solid arrows are dispatch; dashed is RAG return.
Product boundary
One workflow product instead of three overlapping apps. Extension, backend, and dashboard all serve the same user journey: find → tailor → apply → track.
System design
Multi-provider model routing dispatches by question category. pgvector retrieval grounds answers in prior work history. ATS adapters handle Shadow DOM isolation and offline queuing.
Operational proof
40+ backend endpoints, 9-page dashboard, live deployment on Fly.io, 190 automated tests, and 11 ATS adapters covering the major job platforms.
Reusable Engine
tailor-resume
190
automated tests — one hundred ninety pytest cases covering the tailoring engine, verifiable with pytest collect-only.
4
distribution surfaces — four delivery surfaces from one shared core: PyPI package, MCP server hosted on Fly.io, Streamlit web app, and a command-line interface.
MCP
server live on Fly.io — model-context-protocol compatible endpoint discoverable by any MCP-aware agent at tailor-resume-mcp.fly.dev.
The tailoring engine inside AutoApply AI was too useful to keep locked inside one product. Friends wanted it. Agents could call it. So I extracted it properly - 190 tests, four distribution surfaces - and shipped it so any developer can pip install tailor-resume and any MCP-aware agent can call it directly, zero integration work required.
Extracted tailoring core now powers AutoApply AI, an MCP plugin, a browser app, and a CLI - from one honest pipeline. The modularization decision is what created the platform.
Delivery surfaces
Claude Code skill, MCP plugin (Fly.io hosted), Streamlit browser app, CLI, and Python package on PyPI - four surfaces from one shared core.
Quality signal
190 tests and ATS-aware claim discipline signal engineering maturity beyond flashy demos. Parseable by machines, readable by humans.
Discovery Engine
JobScout
130+
career pages monitored — one hundred thirty-plus company-specific scrapers configured in the JobScout target list.
6
ATS platforms — six applicant-tracking-system adapters: Greenhouse, Lever, Workday, Ashby, SmartRecruiters, and iCIMS.
95+
resumes tailored — ninety-five-plus per-role tailored resume PDFs generated from the master profile, each tracked back to the job opening it applied to.
Job boards are noisy by design - most listings are duplicated, aggregated, or irrelevant before you open them. I built JobScout to monitor 130+ company career pages directly, bypassing the middleman, with sponsorship-aware ranking built in. Clean input upstream means better applications downstream.
deployment cycle — release cadence cut from three months to fourteen days by owning CI/CD end-to-end through Azure DevOps; eliminated about eleven weeks of cross-team idle per release.
-67%
CDC ETL compute — CDC ETL runtime cut from thirty minutes to under eight minutes after reengineering full-table reloads into incremental merge-upserts.
~1hr
saved per AAG copy-down — idempotent Azure DevOps pipeline for Contained Always-On Availability Groups (restore plus security plus CDC plus listener validation) eliminated about an hour of manual orchestration on each of twenty-plus daily copy-down requests.
Production data platform work at ExponentHR: CI/CD ownership compressed deployment from 3 months to 14 days, CDC ETL -67% compute cost, payroll-critical AAG database failover automation, idempotent Azure DevOps pipelines.
Azure DevOpsCDC ETLAAG failoverMicrosoft Fabric
ML infrastructure
ML Pipeline Projects
Fraud Detection ML Platform
Streaming fraud scoring with Kafka ingest, FastAPI + LightGBM model serving, MLflow experiment tracking, and Prometheus + Grafana observability. Full model lifecycle in containers.
Real-time risk analytics: Kafka streaming ingest, Spark historical-simulation VaR, FastAPI + Streamlit dashboard for live P&L visibility.
Figure: Portfolio-Risk topology, as it actually runs today. The Spark-to-FastAPI link is currently console-only; FastAPI falls back to synthetic data until the CSV sink lands.
automated tests — three hundred thirty-plus pytest cases passing as of the v1.0 supply-chain hardening release; verifiable with pytest collect-only.
Sigstore + CodeQL
supply-chain hardening — every release signed via Sigstore through GitHub Actions Trusted Publisher OIDC, plus CodeQL static analysis on every pull request; verifiable in the publish and codeql workflow files.
stdlib
only — zero runtime deps — the project dependencies array in pyproject.toml is empty; the package uses Python standard library only at runtime.
Pre/post-commit hooks that snapshot repo context for LLM agents. Sigstore-signed PyPI releases via OIDC Trusted Publisher; CodeQL static analysis on every PR; Dependabot tracks the GitHub Actions ecosystem.
Problem, constraints, design, tradeoffs, outcome. Each post focuses on what didn't work and why.
Supply-chain · 12 min read · Apr 2026
Five Claude agents audited my plugin release in parallel. They caught 41 issues.
A defender-side playbook for shipping a Claude Code plugin with PyPI-tier supply-chain hygiene — Sigstore signing, OIDC Trusted Publisher, CodeQL, property-tested telemetry, and the multi-agent review workflow that found everything I missed.
The LLM cascade as a routing system, not a fallback.
Five providers, circuit breakers per provider, per-category routing, and a reward loop. What worked: Claude-first with cost-aware overflow. What didn't: naive failover by HTTP status code — LLMs fail in much weirder ways than 500s.
Compressing a deploy cycle from 3 months to 14 days.
CI/CD ownership at ExponentHR — Azure DevOps pipelines, idempotent CDC ETL with merge-upserts, AAG copy-down automation. Honest tradeoffs section: which gates I tried to remove and learned to keep.
Two Booleans and a Production Bug: State Design in AutoApply AI
A seemingly simple boolean flag caused a state bug that took three false fixes before the real model emerged. What it taught about representing apply state.
The LLM Cascade as a Routing System, Not a Fallback
Five providers, circuit breakers, per-category routing, and a reward loop. Why the model routing layer is the most interesting engineering surface in AutoApply AI.
"Naren is a great talent in the product ownership space. He is a critical thinker and understands how to analyze a problem, and convert that problem into a tangible requirements doc. Most importantly, he is a quick and humble learner that loves researching the areas of the business that are new to him."
"I had the pleasure of working closely with Narendranath for 3+ years and can confidently attest to his outstanding performance and exceptional work ethic. He had a keen eye for solving business problems by identifying trends to improve the overall business."
SM
Snehal MoniOperations & Policy Analyst, Oregon Health Authority
LinkedIn recommendation
"Narendranath has been an efficient source to the team at all times. He expertly trained himself to adapt to the various sectors of the business and was one of the few people who took up every challenge offered to him."
Based in Dallas TX, open to remote. I'm most useful to teams building production AI systems who want someone who can own the full stack - from retrieval design through deployment and observability.
Response within 24 hours. Based in Dallas TX. Open to Senior AI Platform Engineer, Applied AI, and Backend AI roles.