Senior Data Engineer · Tbilisi, Georgia

Archil
Chachanidze

I build data pipelines that don't break when upstream systems inevitably do — late arrivals, schema drift, partial failures, silent corruption. 7+ years on GCP, 50+ production DAGs, and enough incident postmortems to stop trusting APIs blindly. I also build the tools I wish existed when testing pipelines against chaos.

7+
Years in Data Eng
50+
Production DAGs · ~500 tasks
40%
Failure reduction, idempotency
20+
Chaos fault types built
Core Languages
Python SQL JavaScript
Cloud & GCP
BigQuery Apache Airflow Cloud Functions Cloud Run Pub/Sub Cloud Scheduler Cloud Storage
Data Engineering
ETL / ELT Pipeline Architecture Streaming Schema Design Data Validation Batch Processing Event-driven Arch
Backend & Storage
FastAPI Redis PostgreSQL REST APIs WebSocket Microservices
Infrastructure
Docker Kubernetes Terraform GitHub Actions Prometheus Grafana Linux
Familiar With
Apache Spark dbt Snowflake Databricks Kafka AWS Azure
Dec 2021 – Present
Sweeft Digital / Making Science
Senior Data Engineer
  • Architected pipelines processing 1–5M rows daily across multiple BUs, 99.9% uptime
  • Designed modular framework supporting 50+ production DAGs, 500+ orchestrated tasks
  • Built end-to-end marketing analytics integrating GA4, CM360, Adobe Analytics → BigQuery
  • Reduced pipeline failures 40% via idempotent processing and automated validation
  • Led 1-year migration of 100+ legacy scripts to cloud-native GCP architecture
Oct 2020 – Nov 2021
Olmait
Python Backend Developer
  • Backend services and data integration pipelines using Cloud Functions and Cloud Run
  • Implemented CI/CD pipelines and automated testing for stable production deployments
Nov 2019 – Oct 2020
Various Clients
Freelance Data Engineer
  • Automation tools, ETL pipelines, web scrapers and REST integrations for e-commerce clients
Oct 2018 – Nov 2019
Birtvi
Python Developer / DevOps
  • Automated trading systems integrating multiple crypto exchange APIs with real-time processing
  • Maintained Kubernetes clusters and Dockerized services
Mock Data Engine API

I needed realistic, broken data to test my own pipelines against — late arrivals, corrupted records, schema drift mid-stream. Nothing off the shelf did that, so I built it. Declarative YAML schemas, 20+ injectable fault types, dual-layer persistence, WebSocket streaming. It's a test harness that thinks like an adversary.

20,000+items/sec (WebSocket)
20+chaos fault types
dual-layerRedis + Postgres persistence
FastAPI Redis PostgreSQL Docker Prometheus Grafana Chaos Engineering WebSocket
Marketing Analytics Platform Redesign

Led end-to-end redesign of analytics infrastructure serving 3 cross-functional teams. Unified fragmented data sources into automated Airflow pipelines on GCP.

40%fewer SLA breaches
15 hrs/wkmanual reporting eliminated
Airflow BigQuery GCP GA4 Adobe Analytics