
Data Engineering · Cloud Pipelines · ML-Ready Infrastructure
Ibuilddatasystemsthatscale. Full stop.
Senior Data Engineer and Lead Backend Engineer at the Philadelphia Inquirer. I build data platforms and ETL/ELT pipelines using dbt, Dagster, PySpark, and Airflow, alongside cloud infrastructure and AI/ML systems handling millions of requests daily.
About Me
Most people don't pivot from running a precious metals business to building data platforms that serve a million readers daily. But here we are.
After a decade in strategic communications and business operations, I made a calculated bet during COVID: teach myself full-stack development through a MERN bootcamp, then get my hands dirty with real code. Three and a half years later, I'm architecting cloud infrastructure and data pipelines at the Philadelphia Inquirer, designing systems that ingest, process, and serve content at scale.
My unconventional path gives me something most engineers don't have: deep business context. I understand both the technical requirements and the strategic implications. Whether I'm optimizing database queries, containerizing microservices, or building ML pipelines, I'm solving problems with measurable outcomes: systems that cut costs, reduce latency, and ship on time.
My focus now is data engineering and AI/ML infrastructure: building the pipelines, platforms, and systems that power intelligent applications.

When I'm not engineering systems, you'll find me on the mats training Brazilian Jiu-Jitsu, here with Professor Fred Silva at Lucas Lepri Jiu Jitsu Academy in Charlotte, NC. The discipline, problem-solving, and continuous improvement translate directly to how I approach code.
Tech Stack
Data Engineering
Core stack
Cloud & AI
Platforms & pipelines
What I've Built
Production systems powering digital media at scale, from serverless data pipelines to AI-powered applications over decades of content.
Health & Activity Analytics Pipeline
Designed a production-grade data platform architecture around asset-centric orchestration, then directed AI tooling to build within it. I architected the medallion model (Bronze/Silver/Gold), defined data quality contracts via Great Expectations, and established Kafka streaming patterns. The result: a documented human-in-the-loop workflow where architectural decisions are mine and mechanical implementation is AI-directed. 17 passing dbt tests, 3 validated GX suites, real-time streaming extension via Kafka.
▶ How it works▼ How it works
graph LR; Oura/GitHub/OpenWeatherMap -> Dagster Orchestration -> DuckDB Bronze -> Silver -> Gold -> dbt Core Transforms -> Streamlit Dashboard.
Email Service Provider Migration
The Philadelphia Inquirer needed to exit Salesforce Marketing Cloud without losing subscriber history or segmentation, so I designed a controlled migration path. I built pipelines from Salesforce Service Cloud into Marigold SailThru with rigorous field mapping, validation, and reconciliation. The cutover preserved segmentation and engagement history, achieved zero data loss, and reduced annual platform costs by over $100K.
▶ How it works▼ How it works
graph LR; Salesforce Marketing Cloud -(Legacy State)-> Migration Pipelines; Salesforce Service Cloud -(Subscriber Records)-> Migration Pipelines -> Field Mapping + Validation -> Reconciliation Layer -> Marigold SailThru -> Segmented Lists and Engagement History.
Enterprise production system
Serverless ETL Platform
Legacy billing ETL ran on always-on servers and required frequent manual intervention, so I replaced it with an event-driven AWS architecture. The flow used EventBridge, Lambda, SQS, DynamoDB, and S3 to decouple ingestion, transformation, and reconciliation while maintaining state and auditability. This shift cut infrastructure costs by 75% and reduced manual operations by half.
▶ How it works▼ How it works
graph LR; Billing Event Trigger -> EventBridge -> Lambda Ingest -> SQS Queue -> Lambda Transform -> S3 Raw Storage and DynamoDB State + Audit -> Lambda Reconcile.
Enterprise production system
AI/RAG Historical Archive
The Inquirer's 45-year archive was locked in siloed content systems with no unified search capability, so I designed and deployed a privately-hosted RAG application to surface it. I built an embedding pipeline over 1977-to-present articles, set up a vector retrieval layer, and hosted the inference stack internally to satisfy legal and editorial controls around proprietary content. The result: journalists and editors can now surface decades of institutional knowledge in seconds.
Enterprise production system
ETL Pipeline Reengineering
Legacy Informatica pipelines were brittle, expensive to maintain, and locked behind proprietary licensing that inflated costs without delivering flexibility. I rewrote the critical paths in Python and PySpark, replacing vendor-owned processes with testable, version-controlled code. The migration delivered 50% better resilience on failure recovery and eliminated the Informatica licensing overhead entirely.
Enterprise production system
Internal HR Chatbot
HR was fielding repetitive questions about PTO, org structure, and policy that consumed significant time and created inconsistent answers. I built a RAG-powered chatbot embedded directly in Slack, indexing the employee handbook, org charts, and policy documents into a vector store backed by the OpenAI API. Employees get authoritative, instant answers without leaving their workflow; HR gets fewer tickets.
Enterprise production system
Full Stack GRC Platform
Longevity Consulting's GRC processes were entirely manual, tracked across spreadsheets with no audit trail, no workflow automation, and no visibility across risk domains. I designed and built a full-stack platform with a React front end, Django and FastAPI services, a GraphQL API layer, and containerized deployment on AWS Fargate. The platform gave compliance teams a single source of truth with automated workflows and full audit history.
Enterprise production system
Currently Building
Side projects I ship on nights and weekends. Full-stack execution from idea to production.
Surprise trip planner with PIN-protected reveal pages, itinerary builder, and LemonSqueezy Pro tier. Next.js, Supabase, Vercel.
BJJ lifestyle apparel brand. Pre-launch email capture. Static site with custom WebGL GLSL shader.
Personal AI setup service for Charlotte professionals. OpenClaw configuration, voice calling, and managed care. Next.js, Vercel.
How I Got Here
Senior Data Engineer and Lead Backend Software Engineer
Sep 2022 - PresentThe Philadelphia Inquirer
- ▸Spearheaded ESP migration from Salesforce Marketing Cloud to Marigold SailThru; architected end-to-end data pipelines with Salesforce Service Cloud, delivering one month ahead of schedule, $100K+ annual savings
- ▸Led migration off Informatica ETL by reengineering pipelines in Python/PySpark, 50% more resilient, $120K/year savings
- ▸Architected serverless ETL pipelines on AWS (Serverless Framework) with S3, Lambda, EventBridge, SQS, DynamoDB, replacing legacy subscription/billing systems, enabling real-time Salesforce-to-SailThru sync via event-driven architecture
- ▸Developed privately-hosted AI app enabling chat-based RAG over proprietary historical archive (1977-present)
- ▸Integrated internal HR chatbot with OpenAI API for Slack-based RAG access to handbook, PTO, performance goals, org charts
- ▸Designed real-time streaming pipeline with Apache Kafka for event ingestion into DuckDB analytical warehouse, orchestrated by Dagster sensors for automated data freshness
- ▸Built medallion-architecture data platform (staging, intermediate, marts) with dbt Core, 17+ automated data quality tests, and Great Expectations observability suites
- ▸DevSecOps best practices to minimize Salesforce-induced outages and bot attacks; optimized checkout flow increasing conversion 12% YoY
- ▸Led migration from legacy print subscription management to cloud-first serverless solution, cutting operational costs 75% and manual workload 50%
Software Engineer
Aug 2021 - Sep 2022Longevity Consulting
- ▸Architected full stack GRC platform with React, Django, FastAPI, GraphQL, serverless on AWS FarGate, CICD pipelines
- ▸Led globally distributed agile team through full SDLC
- ▸Developed RPA solutions using UiPath Studio saving $3K+/employee/year
- ▸Assisted US Dept. of Education in Drupal 7 to 9 migration, AWS/GovCloud compliance, PIV/SAML auth
Instructor & Career Coach
Dec 2021 - Dec 2022Nucamp Coding Bootcamp
- ▸Taught 4-5 week cohorts of up to 12 web dev students, 4-hour live workshops weekly
Software Engineer
Jul 2021 - Oct 2022Loxe Inc
- ▸Built decentralized mediation platform on Cardano blockchain
- ▸React + Tailwind frontend, GraphQL, crypto wallet integration (CCVault, Nami, Flint), KYC via SumSub
IT Project Manager
Sep 2011 - Aug 2021Asset Strategies International
- ▸Designed secure services with C# and SQL
- ▸Orchestrated CRM/eCommerce integration (HubSpot, Shopify) with proprietary workflow sync
- ▸Managed database migration from proprietary system
Education & Certifications
UW-Madison · Nucamp · AWS
- ▸BA Public Relations, University of Wisconsin-Madison
- ▸Full Stack Web & Mobile Development Certificate, Nucamp
- ▸Python Data Structures & Algorithms, PostgreSQL with Python, & DevOps Certificate, Nucamp
- ▸AWS Certified Solutions Architect Associate
- ▸AWS Certified Data Engineering Associate
- ▸AWS Certified Cloud Practitioner
Let's Build Something
I'm always interested in challenging data engineering problems, pipeline architecture, and opportunities to build the infrastructure that powers intelligent applications. If you're looking to move off legacy ETL, build a modern data platform, or need someone who translates business requirements into production-grade pipelines, I'd like to hear from you.
Currently based in: Charlotte, NC · Open to: Remote collaboration, consulting, interesting full-time opportunities
Response time: usually same day.