I build reliable, maintainable data pipelines that handle real volume — architected for clarity, configured for flexibility, and built to last beyond the handoff.
Whether you need a pipeline built from scratch, an existing one cleaned up, or a full data warehouse designed and populated, I bring both engineering discipline and domain breadth to the work.
What I deliver:
- Large-scale ETL pipelines — Engineered a config-driven pipeline processing 6M+ US building permit records across multiple municipalities (ConstructIQ), designed so new data sources require zero code changes
- Database design & optimization — PostgreSQL schema design, query optimization, and data modeling for analytical and operational workloads
- Data transformation & enrichment — Cleaning, normalizing, and enriching raw datasets from disparate sources into production-ready formats
- API-connected pipelines — Integrating third-party APIs, LLMs, and external data services into automated ingestion workflows
Core Stack: Python · PostgreSQL · FastAPI · SQLAlchemy · Pandas · Claude API · SerpAPI
I ask about your downstream use case before designing anything upstream. Good pipelines start with good questions.