Resume Snippets

Copyable snippets from my resume, formatted for easy pasting into online job application forms. Each role has multiple formats available - choose whichever fits your needs best.

Title Line

Data Platform Engineer | Stack Technologies Ltd. | Jun 2025 - Present

Role Summary

Lead engineer on the USA data platform rebuild — replacing a legacy Python ETL stack with a production-grade Dagster-native architecture on AWS, with Polars and DuckDB doing the heavy transformation work. Operating at scale: 1.4B+ rows at the mart layer, 400M+ rows in downstream exposures, ~47GB of parquet across landing/intermediate/marts. Operating as an AI-native developer: custom Claude Code skills, sub-agents and hooks act as force multipliers across scaffolding, data inspection, formatting and review — measurably increasing the rate at which new state pipelines and transformation assets ship.

Key Achievements (Bullets)

- Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform.
- Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers.
- Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn.
- Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines.
- Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands.
- Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost.
- Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring.
- Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me.
- Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.

Detailed Highlights (with categories)

Modern Data Platform Architecture:
- Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform.
- Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers.
- Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn.
AI-native Developer Workflow:
- Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines.
- Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands.
Cloud Infrastructure & DevOps:
- Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost.
- Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring.
Framework & Team Enablement:
- Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me.
- Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.

Full Description (Summary + Achievements)

Lead engineer on the USA data platform rebuild — replacing a legacy Python ETL stack with a production-grade Dagster-native architecture on AWS, with Polars and DuckDB doing the heavy transformation work. Operating at scale: 1.4B+ rows at the mart layer, 400M+ rows in downstream exposures, ~47GB of parquet across landing/intermediate/marts. Operating as an AI-native developer: custom Claude Code skills, sub-agents and hooks act as force multipliers across scaffolding, data inspection, formatting and review — measurably increasing the rate at which new state pipelines and transformation assets ship.

- Architected production Dagster platform across 15+ US states and 400+ transformation assets (1.4B+ mart rows, 400M+ exposure rows), replacing a legacy Python ETL stack with a schema-contracted, testable modern data platform.
- Cut processing time on critical datasets through targeted Polars/DuckDB rewrites, incremental materializations, and coalesced fact/mart layers.
- Designed reusable exposure/coalesce patterns so legacy mart schemas ride on top of the new fact-table architecture without consumer churn.
- Built a project-specific Claude Code toolchain — sub-agent, slash commands, auto-format hook — that measurably increases the rate of shipping new pipelines.
- Profiled my own Claude Code usage to surface high-frequency patterns (65% parquet inspection) and promoted them into first-class agents/commands.
- Deployed the platform's AWS footprint with Terraform IaC across staging and production, migrating Batch workloads to ECS for efficiency and cost.
- Implemented CI/CD (GitHub Actions, ruff, pyright, asset schema-contract tests) and the branching/deployment automation that makes staging → prod boring.
- Standardized 100+ Dagster assets with reusable framework patterns and wrote the docs so the team scales beyond me.
- Raised the engineering bar: schema contracts, type-safe Python, architectural docs, and review workflows befitting an enterprise-grade data platform.

Education Snippets

Skills Summary

All Skills (Grouped by Type)

Data & Software:
- AI Engineering & Agentic Development: Claude Code (agents, skills, hooks, slash commands, sub-agents), MCP servers & clients, Anthropic / OpenAI / Perplexity APIs, Vercel AI SDK, Structured outputs & schema-adherent LLM calls, Prompt caching & evals, RAG, Agent frameworks (autogen, crew), LLM-in-the-loop data pipelines, Editors: Claude Code, Cursor, Windsurf
- Data Engineering & Modern Data Stack: dbt (Core & Cloud), Dagster, Snowflake, Snowpark, Polars, DuckDB, Tableau Prep, cube.js & dbt semantic layers, pandas, SQL, Schema contracts
- Cloud, Infra & DevOps: AWS (ECS, RDS, S3), Terraform, Docker, GitHub Actions CI/CD, Vercel, Railway, Supabase, Doppler, pre-commit / ruff / pyright
- Data Visualization & BI: Tableau, PowerBI, Spotfire, Plotly, D3.js, matplotlib, Seaborn, Dash, Bokeh, ggplot2
- Machine Learning & NLP: scikit-learn, XGBoost, PyTorch, TensorFlow / Keras, Hugging Face, nltk, Time-series forecasting, Classification, Anomaly detection
- Databases: Snowflake, PostgreSQL, DuckDB, Neo4j, networkX, MongoDB, SQL Server, Oracle, Hive
- Web App Development: Next.js, React, Astro, Directus (headless CMS), Algolia, Node.js, Express, Tailwind CSS, Playwright, Streamlit, Flask, Dash, Turborepo / npm workspaces
- ML Ops: Snowflake & Snowpark, DataRobot, Streamlit labeling apps, Active learning workflows

Programming Languages:
- Programming Languages: Python, SQL, TypeScript, JavaScript, Bash, R

Oil & Gas:
- Upstream Oil & Gas: Exploitation / Development Engineering, Production Engineering, Operations, Corporate reporting, A&D evaluation
- Petroleum Reserves Evaluation: COGEH, NI 51-101