On-Site Pune Data Engineer

Замовник: AI | Опубліковано: 11.02.2026

I run a data platform built on Spark and Python, and I need an experienced engineer who can sit with me in Pune three to four times a week (hardly 2 hours in a day) to keep it running smoothly. Most of the immediate work revolves around tracking down bugs in existing PySpark jobs, but the role naturally extends to writing fresh code when gaps appear, tightening our data-pipeline orchestration, and mapping end-to-end data lineage so every downstream consumer stays confident in their numbers. Typical day-to-day work you will tackle: • Debug production PySpark jobs and accompanying Python utilities • Refactor or rewrite modules where quick fixes will not suffice • Optimise and monitor pipeline schedules (Palantir Foundry) • Document lineage and hand off clear, reproducible notebooks or scripts Presence in Pune is mandatory. We meet in person at my place near Kharadi for 2 hours in a day three times each week; remote-only applicants will be ignored. Deliverables will be reviewed functionally in the cluster and via code review; payment milestones align with each successfully repaired or newly delivered pipeline component. If this rhythm and tech stack suit you, send across a short note highlighting a comparable debugging challenge you cracked in PySpark, and let’s schedule our first on-site session.