City of London, London, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
Be Doing You'll be a key contributor to the development of a next-generation data platform, with responsibilities including: Designing and implementing scalable data pipelines using Python and Apache Spark Building and orchestrating workflows using AWS services such as Glue , Lambda , S3 , and EMR Serverless Applying best practices in software engineering: CI/CD , version control , automated testing … and modular design Supporting the development of a lakehouse architecture using ApacheIceberg Collaborating with product and business teams to deliver data-driven solutions Embedding observability and quality checks into data workflows Participating in code reviews, pair programming, and architectural discussions Gaining domain knowledge in financial data and sharing insights with the team What They're Looking For … for experience with type hints, linters, and testing frameworks like pytest) Solid understanding of data engineering fundamentals: ETL/ELT, schema evolution, batch processing Experience or strong interest in Apache Spark for distributed data processing Familiarity with AWS data tools (e.g., S3, Glue, Lambda, EMR) Strong communication skills and a collaborative mindset Comfortable working in Agile environments and engaging More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
Be Doing You'll be a key contributor to the development of a next-generation data platform, with responsibilities including: Designing and implementing scalable data pipelines using Python and Apache Spark Building and orchestrating workflows using AWS services such as Glue , Lambda , S3 , and EMR Serverless Applying best practices in software engineering: CI/CD , version control , automated testing … and modular design Supporting the development of a lakehouse architecture using ApacheIceberg Collaborating with product and business teams to deliver data-driven solutions Embedding observability and quality checks into data workflows Participating in code reviews, pair programming, and architectural discussions Gaining domain knowledge in financial data and sharing insights with the team What They're Looking For … for experience with type hints, linters, and testing frameworks like pytest) Solid understanding of data engineering fundamentals: ETL/ELT, schema evolution, batch processing Experience or strong interest in Apache Spark for distributed data processing Familiarity with AWS data tools (e.g., S3, Glue, Lambda, EMR) Strong communication skills and a collaborative mindset Comfortable working in Agile environments and engaging More ❯
able to work across full data cycle. - Proven Experience working with AWS data technologies (S3, Redshift, Glue, Lambda, Lake formation, Cloud Formation), GitHub, CI/CD - Coding experience in Apache Spark, Iceberg or Python (Pandas) - Experience in change and release management. - Experience in Database Warehouse design and data modelling - Experience managing Data Migration projects. - Cloud data platform development … the AWS services like Redshift, Lambda,S3,Step Functions, Batch, Cloud formation, Lake Formation, Code Build, CI/CD, GitHub, IAM, SQS, SNS, Aurora DB - Good experience with DBT, ApacheIceberg, Docker, Microsoft BI stack (nice to have) - Experience in data warehouse design (Kimball and lake house, medallion and data vault) is a definite preference as is knowledge More ❯
MySQL Exposure to Docker, Kubernetes, AWS, Helm, Terraform, Vault, Grafana, ELK Stack, New Relic Relevant experience in the maintenance of data APIs and data lake architectures, including experience with ApacheIceberg, Trino/Presto, Clickhouse, Snowflake, BigQuery. Master's degree in Computer Science or Engineering-related field Equal Opportunity Employer As an Equal Opportunity Employer, qualified applicants will More ❯
Terraform and Kubernetes is a plus! A genuine excitement for significantly scaling large data systems Technologies we use (experience not required): AWS serverless architectures Kubernetes Spark Flink Databricks Parquet. Iceberg, Delta lake, Paimon Terraform Github including Github Actions Java PostgreSQL About Chainalysis Blockchain technology is powering a growing wave of innovation. Businesses and governments around the world are using More ❯
innovation through advanced analytics and research-based problem solving. To be successful you should have: 10 years hands-on experience in AWS data engineering technologies, including Glue, PySpark, Athena, Iceberg, Databricks, Lake Formation, and other standard data engineering tools. Previous experience in implementing best practices for data engineering, including data governance, data quality, and data security. Proficiency in data More ❯
innovation through advanced analytics and research-based problem solving. To be successful you should have: 10 years hands-on experience in AWS data engineering technologies, including Glue, PySpark, Athena, Iceberg, Databricks, Lake Formation, and other standard data engineering tools. Previous experience in implementing best practices for data engineering, including data governance, data quality, and data security. Proficiency in data More ❯
sets. Collaborate with data scientists to deploy machine learning models. Contribute to strategy, planning, and continuous improvement. Required Experience: Hands-on experience with AWS data tools: Glue, PySpark, Athena, Iceberg, Lake Formation. Strong Python and SQL skills for data processing and analysis. Deep understanding of data governance, quality, and security. Knowledge of market data and its business applications. Desirable More ❯
sets. Collaborate with data scientists to deploy machine learning models. Contribute to strategy, planning, and continuous improvement. Required Experience: Hands-on experience with AWS data tools: Glue, PySpark, Athena, Iceberg, Lake Formation. Strong Python and SQL skills for data processing and analysis. Deep understanding of data governance, quality, and security. Knowledge of market data and its business applications. Desirable More ❯
Manchester, Lancashire, England, United Kingdom Hybrid / WFH Options
Lorien
passionate about building scalable, cloud-native data platforms. You'll be a key player in a growing team, helping to shape the future of data infrastructure using AWS, PySpark, Iceberg, and more. From designing high-performance pipelines to supporting a full-scale migration from SQL Server to AWS, this role offers the chance to work on real-time data More ❯
Milton Keynes, Buckinghamshire, South East, United Kingdom
Upbeat Ideas UK Ltd
data sources. Document test results and provide detailed reports on data quality findings. Required Skills and Experience: Proven experience with AWS services such as EMR, Lambda, Redshift, Firehose, S3, Iceberg, Athena, and DynamoDB. Strong understanding of data ingestion, parsing, aggregation, and schema validation processes. Proficiency in SQL for data querying and validation. Experience with Python for scripting and automation. More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Robert Half
Glue Glue Data Catalog EMR services API Gateway Relational database experience with Aurora Postgres (including query performance tuning). Spark experience, including pipelines using Spark on data stored in iceberg table format in S3 and federated through Aurora Postgres. Proven experience with data modeling, medallion/lakehouse architectures, and financial domain familiarity. Ability to solve problems independently and deliver More ❯
vision and roadmap for AI infrastructure and data engineering. Lead, mentor, and scale global engineering teams. Oversee large-scale distributed compute, storage, and streaming systems (e.g., Spark, Kafka, Flink, Iceberg/Delta Lakes). Collaborate with cross-functional teams (Data Science, ML Engineering, Product). Build and maintain pipelines, governance frameworks, dashboards, and analytics systems. Ensure reliability, scalability, compliance More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Robert Half
manage data solutions on AWS using serverless technologies such as Lambda, Glue, Glue Data Catalog, EMR serverless, and API Gateway. Implement and manage large-scale data processing with Spark (Iceberg tables in S3, Gold layer in Aurora Postgres). Collaborate with data scientists, analysts, and stakeholders to ensure data availability, accessibility, and reliability. Design and maintain robust data models … particularly with serverless data engineering tools (Lambda, Glue, Glue Data Catalog, EMR serverless, API Gateway, S3). Solid Spark experience with large-scale data pipelines and data lakehouse architectures (Iceberg format a plus). Hands-on experience with data modelling and designing scalable data architectures. Strong understanding of the financial domain (preferred but not mandatory). Excellent problem-solving More ❯