City of London, London, United Kingdom Hybrid / WFH Options
Rise Technical Recruitment Limited
become a trusted partner across a wide range of businesses. In this role you'll take ownership of the reliability and performance of large-scale date pipelines built on AWS, Apache Flink, Kafka, and Python. You'll play a key role in diagnosing incidents, optimising system behaviour, and ensuring reporting data is delivered on time and without failure. The … ideal candidate will have a strong experience working with streaming and batch data systems, a solid understanding of monitoring a observability, and hands-on experience working with AWS, Apache Flink, Kafka, and Python. This is a fantastic opportunity to step into a SRE role focused on data reliability in a modern cloud native environment, with full ownership of … incident management, architecture, and performance. The Role: *Maintaining and monitoring real-time and batch data pipelines using Flink, Kafka, Python, and AWS *Act as an escalation point for critical data incidents and lead root cause analysis *Optimising system performance, define SLIs/SLOs, and drive reliability *Woking closely with various other departments and teams to architect scalable, fault-tolerant More ❯
will help drive the evolution of our data architecture as we move from Redshift to Snowflake. Looking for someone with extensive experience with cloud providers? Hands-on experience with AWS services such as Glue (Spark), Lambda, StepFunctions, ECS, Redshift, and SageMaker. Looking for someone with hands-on development Conducting code reviews, mentoring through pair programming. Looking … Building APIs, integrating with microservices, or contributing to backend systems — not just data pipelines or data modelling. CI/CD and Infrastructure-as-Code Tools like GitHub Actions, Jenkins, AWS CDK, CloudFormation, Terraform. Key Responsibilities: Design and implement scalable, secure, and cost-efficient data solutions on AWS, leveraging services such as Glue, Lambda, S3, Redshift, and Step … higher in a technical discipline Proven experience as a data engineer with strong hands-on programming skills and software engineering fundamentals, with experience building scalable solutions in cloud environments (AWS preferred) Extensive experience in AWS services, e.g. EC2, S3, RDS, DynamoDB, Redshift, Lambda, API Gateway Solid foundation in software engineering principles, including version control (Git), testing, CI/ More ❯
processes to improve efficiency, scalability, reliability and observability. Drive Engineering Excellence: Lead and manage all engineering activities across internal and external teams, ensuring high productivity and quality of execution. AWS Expertise: Strong expertise across AWS products, including S3, Glue, Spark, DBT, Terraform, and Redshift. Roadmap Prioritisation: Prioritize and manage engineering activities and personnel to deliver on a roadmap … Skills for the Data Operations Manager: Technology Degree with at least 5 years’ experience in data Proven experience in managing engineering teams in a fast-paced environment. Knowledge of AWS services and tools, including S3, StepFunctions, Spark, DBT, Terraform, and Redshift. Strong leadership and communication skills, with the ability to inspire and motivate a diverse team. More ❯