of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Solirius Reply
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
communication skills We're excited if you have 7+ years of experience delivering multi tier, highly scalable, distributed web applications Experience working with Distributed computing frameworks knowledge: Hive/Hadoop, Apache Spark, Kafka, Airflow Working with programming languages Python , Java, SQL. Working on building ETL (Extraction Transformation and Loading) solution using PySpark Experience in SQL/NoSQL database design More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka, Gurobi, or MySQL) 4+ years of experience designing, building and optimizing data pipelines and ETL workflows at scale 4+ years of experience with UNIX/Linux More ❯
Elizabeth, New Jersey, United States Hybrid/Remote Options
ALTA IT Services
of experience may be considered in lieu of a degree. Proficiency in programming languages like Python or Java, strong SQL skills, and knowledge of big data tools like ApacheHadoop, Spark, or Kafka. Experience with cloud platforms (AWS, Azure, GCP) and data warehousing solutions (Snowflake, Redshift, BigQuery) Self-driven and have demonstrated the ability to work independently with minimum More ❯
Luton, England, United Kingdom Hybrid/Remote Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
London, South East, England, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with Relational DBs (MySQL, Postges, SQL Server More ❯
and troubleshooting application deployments using CI/CD pipelines (Azure DevOps Tool Suites along with Azure Pipelines and Portal) 9+ years of experience with the following tools and technologies: Hadoop, Spark, Kafka, Nifi, Relational SQL or NoSQL databases 5+ years of Technical leadership experience guiding development teams to cloud Experience with these tools; PowerBI, Azure DevOps, SaaS, Data Integration More ❯
to support analytical and business goals. Monitor, troubleshoot , and enhance data performance and infrastructure. Key Skills & Experience: Strong experience with SQL/NoSQL databases, data warehousing, and big data (Hadoop, Spark). Proficient in Python, Java, or Scala with solid OOP and design pattern understanding. Expertise in ETL tools, DevOps and orchestration frameworks (Airflow, Apache NiFi). Hands-on More ❯
in data engineering, with a strong emphasis on data design and architecture. Proven proficiency in SQL and experience with relational databases. Practical experience with big data technologies such as Hadoop or Spark. In-depth understanding of data warehousing concepts and ETL frameworks. Familiarity with cloud platforms including AWS, Azure, or GCP. Strong analytical and problem-solving skills, with the More ❯
Azure Data FactoryExperience in Unix/Linux environments and shell scripting.Excellent problem-solving, analytical, and communication skills. Preferred SkillsExperience with Informatica Intelligent Cloud Services (IICSExposure to big data platforms (Hadoop, Spark) or data lake architectures.Familiarity with Agile or Scrum methodologies.Background in finance, healthcare, or retail data environments is a plus.Experience integrating data from cloud platforms (AWS, Azure, GCP More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Areti Group | B Corp™
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯