of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
communication skills We're excited if you have 7+ years of experience delivering multi tier, highly scalable, distributed web applications Experience working with Distributed computing frameworks knowledge: Hive/Hadoop, Apache Spark, Kafka, Airflow Working with programming languages Python , Java, SQL. Working on building ETL (Extraction Transformation and Loading) solution using PySpark Experience in SQL/NoSQL database design More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
providers, preferably Azure (AWS/GCP also valued). Experience with modern data lakehouse concepts and architectures. Experience ingesting and transforming data from sources such as Kafka, REST APIs, Hadoop, etc. Experience with data pipeline orchestration tools (Azure Data Factor or similar). Experience implementing ETL/ELT pipelines using modern data lakehouse solutions. Experience with data quality, observability More ❯
analytical thinking capabilities. Experience working in enterprise environments with cross-functional teams. Preferred Experience Experience with cloud data environments (AWS, Azure, or Google Cloud Familiarity with big data platforms (Hadoop, Spark, Databricks Background in industries such as finance, healthcare, government, or manufacturing. Experience with model governance, validation, and lifecycle management. Education Bachelor's or Master's degree in Data More ❯
London, South East, England, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯