Resume Snippets
Copyable snippets from my resume, formatted for easy pasting into online job application forms. Each role has multiple formats available - choose whichever fits your needs best.
Title Line
Data Platform Engineer | Stack Technologies Ltd. | Jun 2025 - Present
Role Summary
Lead engineer on the USA data platform rebuild — replacing a legacy Python ETL stack with a production-grade Dagster-native architecture on AWS, with Polars and DuckDB doing the heavy transformation work. Operating at scale: 1.4B+ rows at the mart layer, 400M+ rows in downstream exposures, ~47GB of parquet across landing/intermediate/marts. Operating as an AI-native developer: custom Claude Code skills, sub-agents and hooks act as force multipliers across scaffolding, data inspection, formatting and review — measurably increasing the rate at which new state pipelines and transformation assets ship.
Key Achievements (Bullets)
- Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform. - Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers. - Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn. - Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines. - Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands. - Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost. - Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring. - Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me. - Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.
Detailed Highlights (with categories)
Modern Data Platform Architecture: - Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform. - Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers. - Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn. AI-native Developer Workflow: - Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines. - Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands. Cloud Infrastructure & DevOps: - Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost. - Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring. Framework & Team Enablement: - Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me. - Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.
Full Description (Summary + Achievements)
Lead engineer on the USA data platform rebuild — replacing a legacy Python ETL stack with a production-grade Dagster-native architecture on AWS, with Polars and DuckDB doing the heavy transformation work. Operating at scale: 1.4B+ rows at the mart layer, 400M+ rows in downstream exposures, ~47GB of parquet across landing/intermediate/marts. Operating as an AI-native developer: custom Claude Code skills, sub-agents and hooks act as force multipliers across scaffolding, data inspection, formatting and review — measurably increasing the rate at which new state pipelines and transformation assets ship. - Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform. - Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers. - Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn. - Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines. - Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands. - Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost. - Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring. - Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me. - Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.
Education Snippets
Skills Summary
All Skills (Grouped by Type)
Data & Software: - AI Engineering & Agentic Development: Claude Code (agents, skills, hooks, slash commands, sub-agents), MCP servers & clients, Anthropic / OpenAI / Perplexity APIs, Vercel AI SDK, Structured outputs & schema-adherent LLM calls, Prompt caching & evals, RAG, Agent frameworks (autogen, crew), LLM-in-the-loop data pipelines, Editors: Claude Code, Cursor, Windsurf - Data Engineering & Modern Data Stack: dbt (Core & Cloud), Dagster, Snowflake, Snowpark, Polars, DuckDB, Tableau Prep, cube.js & dbt semantic layers, pandas, SQL, Schema contracts - Cloud, Infra & DevOps: AWS (ECS, RDS, S3), Terraform, Docker, GitHub Actions CI/CD, Vercel, Railway, Supabase, Doppler, pre-commit / ruff / pyright - Data Visualization & BI: Tableau, PowerBI, Spotfire, Plotly, D3.js, matplotlib, Seaborn, Dash, Bokeh, ggplot2 - Machine Learning & NLP: scikit-learn, XGBoost, PyTorch, TensorFlow / Keras, Hugging Face, nltk, Time-series forecasting, Classification, Anomaly detection - Databases: Snowflake, PostgreSQL, DuckDB, Neo4j, networkX, MongoDB, SQL Server, Oracle, Hive - Web App Development: Next.js, React, Astro, Directus (headless CMS), Algolia, Node.js, Express, Tailwind CSS, Playwright, Streamlit, Flask, Dash, Turborepo / npm workspaces - ML Ops: Snowflake & Snowpark, DataRobot, Streamlit labeling apps, Active learning workflows Programming Languages: - Programming Languages: Python, SQL, TypeScript, JavaScript, Bash, R Oil & Gas: - Upstream Oil & Gas: Exploitation / Development Engineering, Production Engineering, Operations, Corporate reporting, A&D evaluation - Petroleum Reserves Evaluation: COGEH, NI 51-101