of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
data modelling tools, data warehousing, ETL processes, and data integration techniques. Experience with at least one cloud data platform (e.g. AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark). Strong knowledge of data workflow solutions like Azure Data Factory, Apache NiFi, Apache Airflow etc Good knowledge of stream and batch processing solutions like Apache Flink, ApacheMore ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
entity-relationship modeling, and dimensional data modeling . Experience with relational and NoSQL databases such as Oracle, Sybase, PostgreSQL, SQL Server, MongoDB . Familiarity with big data platforms (e.g., Hadoop, Snowflake). Prior experience with ETL tools or as a SQL developer . Proficiency in Python for data engineering and Tableau for reporting and dashboards. Exposure to cloud platforms More ❯
data modelling tools, data warehousing, ETL processes, and data integration techniques. Experience with at least one cloud data platform (e.g. AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark). Strong knowledge of data workflow solutions like Azure Data Factory, Apache NiFi, Apache Airflow Good knowledge of stream and batch processing solutions like Apache Flink, Apache Kafka More ❯
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
Luton, England, United Kingdom Hybrid / WFH Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
luton, bedfordshire, east anglia, united kingdom Hybrid / WFH Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
bedford, east anglia, united kingdom Hybrid / WFH Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
watford, hertfordshire, east anglia, united kingdom Hybrid / WFH Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
Azure Data FactoryExperience in Unix/Linux environments and shell scripting.Excellent problem-solving, analytical, and communication skills. Preferred SkillsExperience with Informatica Intelligent Cloud Services (IICSExposure to big data platforms (Hadoop, Spark) or data lake architectures.Familiarity with Agile or Scrum methodologies.Background in finance, healthcare, or retail data environments is a plus.Experience integrating data from cloud platforms (AWS, Azure, GCP More ❯
Manchester, Lancashire, United Kingdom Hybrid / WFH Options
CHEP UK Ltd
such as Python, R, and SQL for data analysis and model development. Experience working with cloud computing platforms including AWS and Azure, and familiarity with distributed computing frameworks like Hadoop and Spark. Deep understanding of supply chain operations and the ability to apply data science methods to solve real-world business problems effectively. Strong foundational knowledge in mathematics and More ❯
Flows, Conduct>It, Express>It, Metadata Hub, and PDL. Hands-on experience with SQL , Unix/Linux shell scripting , and data warehouse concepts . Familiarity with big data ecosystems (Hadoop, Hive, Spark) and cloud platforms (AWS, Azure, GCP) is a plus. Proven ability to troubleshoot complex ETL jobs and resolve performance issues. Experience working with large-scale datasets and More ❯
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move Mission & impact: Work on projects where data-driven decisions have real More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Areti Group | B Corp™
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move Mission & impact: Work on projects where data-driven decisions have real More ❯
london (city of london), south east england, united kingdom Hybrid / WFH Options
Areti Group | B Corp
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move Mission & impact: Work on projects where data-driven decisions have real More ❯