Data Engineer
Are you a Data Engineer who enjoys building production-grade pipelines, optimising performance, and working with modern Python tooling (DuckDB/Polars) on time-series datasets?
I'm supporting a UK-based fintech in their search for a hands-on Python Data Engineer to help build and improve the data infrastructure powering a unified data + analytics API for financial markets participants.
You'll sit in a engineering/analytics team and take ownership of pipelines end-to-end — from onboarding new datasets through to reliability, monitoring and data quality in production.
In this role, you'll:
- Build, streamline and improve ETL/data pipelines (prototype → production)
- Ingest and normalise high-velocity time-series datasets from multiple external sources
- Work heavily in Python with a modern stack including DuckDB and Polars (plus Parquet/PyArrow)
- Orchestrate workflows and improve reliability (they use Temporal — similar orchestration experience is fine)
- Improve data integrity and visibility: validations, automated checks, backfills, monitoring/alerting
- Support downstream analytics and client-facing outputs (dashboards/PDF/Plotly — least important)
What's in it for you?
- Modern data stack – DuckDB/Polars + Parquet/Arrow in a genuinely hands-on environment
- Ownership & impact – You'll be close to the data flows and have real influence on performance and reliability
- Market data exposure – Work with complex financial datasets (experience helpful, interest is enough)
- Hybrid London – London preferred, with 2–3 days in the office
- Start ASAP – Interviewing now
What my client is looking for:
- Strong Python + SQL fundamentals (data engineering / ETL / pipeline ownership)
- Hands-on experience with DuckDB and/or Polars (DuckDB especially valuable)
- Experience operating pipelines in production (monitoring, backfills, incident/RCA mindset, data quality)
- Cloud experience with demonstrable production use (Azure preferred)
- Clear communicator, comfortable working across engineering/analytics stakeholders
Nice to have:
- Time-series data experience (market data, telemetry, pricing, events)
- Streaming exposure (Kafka/Event Hubs/Kinesis)
- Experience with Temporal (or similar orchestrators like Airflow/Dagster/Prefect)
- Any exposure to AI agents / automation tooling
Apply now!