Data Engineer Azure Data Platform

About Synextra Synextra is a Microsoft-specialist Managed Service Provider headquartered in Warrington, operating as a premium partner to regulated mid-market organisations including law firms, financial services firms, and mortgage lenders. We're deliberately small - around 35 people - because we believe the best outcomes come from technical depth, not headcount. Our AI Services Division is growing fast, and we're building out a serious data and engineering capability to match. This is a chance to get in early and shape how that function operates.

The Role We're looking for a technically driven Azure Data Engineer to join our data platform team. You'll design, build, and maintain production-grade data pipelines on Microsoft Azure - transforming complex, diverse datasets into analytics-ready formats that power business intelligence and AI initiatives for our clients and internally.The ideal candidate treats pipelines and infrastructure as code, with a genuine passion for software engineering in a data context. You'll work across the modern Azure data stack - ADF, ADLS Gen2, PySpark, Delta Lake - with increasing exposure to Microsoft Fabric as the platform matures. You'll collaborate closely with customers and internal teams to ensure data is structured and governed for reliable downstream consumption.This is a hands-on engineering role with room to grow into leadership: you'll champion DevOps best practices, contribute to architectural decisions, and help mentor junior engineers as the team scales.

Responsibilities

  • Architect and write production-grade ELT/ETL data pipelines using PySpark and Python within Azure ecosystem.

  • Build custom, reusable data processing frameworks and libraries in Python/Scala to streamline ingestion and transformation tasks across the engineering team

  • Programmatically ingest large volumes of structured and unstructured data from REST APIs, streaming platforms (e.g. Event Hubs, Kafka), and legacy databases into ADLS Gen2 and OneLake

  • Develop structured data models aligned to Lakehouse, Medallion Architecture, and Delta Lake patterns

  • Continuously profile, debug, and optimise Spark jobs, SQL queries, and Python scripts for maximum performance and cost-efficiency at scale

  • Champion DevOps best practices: implement infrastructure-as-code (Terraform), automated testing, and CI/CD deployment pipelines via Git and Azure DevOps

  • Identifying patterns in recurring issues and engineering permanent solutions

  • Write comprehensive unit and integration tests for all data pipelines to ensure data integrity; enforce data governance protocols, RBAC, and encryption standards across all environments

Requirements

Essential Technical Skills

  • Advanced proficiency in Python and PySpark, writing clean, modular, object-oriented code for data transformations

  • Strong command of SQL (T-SQL, Spark SQL) for data exploration, validation, and final-stage modelling

  • Deep hands-on experience with Microsoft Fabric and its tooling such as Azure Data Factory (ADF), and Azure Data Lake Storage (ADLS Gen2)

  • Practical experience with Git, branching strategies, automated testing (e.g. pytest), and CI/CD orchestration via Azure DevOps

  • Proven commercial track record of deploying complex data solutions on the Microsoft Azure platform

  • Experience collaborating with a range of stakeholders to structure data for downstream consumption (e.g. MLflow, Power BI semantic models)

  • Infrastructure-as-code experience with Terraform for Azure resource provisioning

Desirable Technical Skills

  • Familiarity with streaming data architectures (Spark Structured Streaming)

  • Knowledge of complementary modern data stack tools such as dbt for SQL-based transformations

  • Experience integrating Large Language Models (LLMs) or operationalising AI/ML models

Personal Qualities

  • Exceptional problem-solving abilities and a persistent, detail-oriented approach to debugging complex code

  • Strong communication skills to effectively translate business requirements into technical architectures

  • A proactive mindset focused on continuous learning and staying ahead of the rapidly evolving data landscape

  • Willingness to review code submissions, enforce coding standards, and mentor junior engineers on the team

Preferred Background

  • 3–5+ years in software engineering, data engineering, or Big Data environments with a code-first approach

  • Proven commercial experience deploying and maintaining complex data solutions on Microsoft Azure

  • Experience working in cross-functional teams

Job Details

Company
Synextra
Location
Birchwood Park, Birchwood, Warrington, Cheshire, England, United Kingdom
Employment Type
Full-Time
Salary
£55,000 - £70,000 per annum
Posted