of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
quality-driven mindset Experience with CI/CD pipelines and writing complex queries Excellent communication and collaboration in fast-paced teams Bonus points for: Experience with big data technologies (Hadoop, Spark, Kafka) Knowledge of best practices in coding, data versioning, monitoring, and optimisation Understanding of cloud data services and distributed/parallel computing Workflow scheduling/monitoring experience Relevant More ❯
data modelling tools, data warehousing, ETL processes, and data integration techniques. · Experience with at least one cloud data platform (e.g. AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark). · Strong knowledge of data workflow solutions like Azure Data Factory, Apache NiFi, Apache Airflow etc · Good knowledge of stream and batch processing solutions like Apache Flink, ApacheMore ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Reporting tools (e.g. Tableau, PowerBI, Qlik) GDPR and Government More ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Solirius Reply
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
Luton, England, United Kingdom Hybrid/Remote Options
easyJet
Understanding of ML development workflow and knowledge of when and how to use dedicated hardware. Significant experience with Apache Spark or any other distributed data programming frameworks (e.g. Flink, Hadoop, Beam) Familiarity with Databricks as a data and AI platform or the Lakehouse Architecture. Experience with data quality and/or and data lineage frameworks like Great Expectations, dbt More ❯
communication skills We're excited if you have 7+ years of experience delivering multi tier, highly scalable, distributed web applications Experience working with Distributed computing frameworks knowledge: Hive/Hadoop, Apache Spark, Kafka, Airflow Working with programming languages Python , Java, SQL. Working on building ETL (Extraction Transformation and Loading) solution using PySpark Experience in SQL/NoSQL database design More ❯
to support analytical and business goals. Monitor, troubleshoot , and enhance data performance and infrastructure. Key Skills & Experience: Strong experience with SQL/NoSQL databases, data warehousing, and big data (Hadoop, Spark). Proficient in Python, Java, or Scala with solid OOP and design pattern understanding. Expertise in ETL tools, DevOps and orchestration frameworks (Airflow, Apache NiFi). Hands-on More ❯
in data engineering, with a strong emphasis on data design and architecture. Proven proficiency in SQL and experience with relational databases. Practical experience with big data technologies such as Hadoop or Spark. In-depth understanding of data warehousing concepts and ETL frameworks. Familiarity with cloud platforms including AWS, Azure, or GCP. Strong analytical and problem-solving skills, with the More ❯
London, South East, England, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Tenth Revolution Group
problem-solving skills, and the ability to think critically and analytically High experience in documentation and data dictionaries Knowledge of big data technologies and distributed computing frameworks such as Hadoop and Spark Excellent communication skills to effectively collaborate with cross-functional teams and present insights to business stakeholders Please can you send me a copy of your CV if More ❯
Crewe, Cheshire, United Kingdom Hybrid/Remote Options
Manchester Digital
AI model development. Expertise in Python, R, or Julia, with proficiency in pandas, NumPy, SciPy, scikit-learn, TensorFlow, or PyTorch. Experience with SQL, NoSQL, and big data technologies (Spark, Hadoop, Snowflake, Databricks, etc.). Strong background in statistical modelling, probability theory, and mathematical optimization. Experience deploying machine learning models to production (MLOps, Docker, Kubernetes, etc.). Familiarity with AWS More ❯
Manchester, Lancashire, United Kingdom Hybrid/Remote Options
CHEP UK Ltd
such as Python, R, and SQL for data analysis and model development. Experience working with cloud computing platforms including AWS and Azure, and familiarity with distributed computing frameworks like Hadoop and Spark. Deep understanding of supply chain operations and the ability to apply data science methods to solve real-world business problems effectively. Strong foundational knowledge in mathematics and More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Areti Group | B Corp™
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
Security project experience. Experience with Palantir Foundry (full training provided). Familiarity with AI/ML Ops pipelines , real-time analytics, or edge deployments. Big Data stack knowledge (e.g., Hadoop, Spark, Kafka). GenAI/LLM experience (e.g., AWS Bedrock, LangChain). Why this is a great move 🌳 Mission & impact: Work on projects where data-driven decisions have real More ❯
Flows, Conduct>It, Express>It, Metadata Hub, and PDL. Hands-on experience with SQL , Unix/Linux shell scripting , and data warehouse concepts . Familiarity with big data ecosystems (Hadoop, Hive, Spark) and cloud platforms (AWS, Azure, GCP) is a plus. Proven ability to troubleshoot complex ETL jobs and resolve performance issues. Experience working with large-scale datasets and More ❯
Information Systems, or a related discipline. Desirable Experience Background or internship experience within financial services or technology. Exposure to Java. Experience managing on-premise or hybrid data infrastructure (e.g. Hadoop). Knowledge of workflow orchestration tools such as Apache Airflow. Postgraduate degree in Computer Science, Data Science, or related field. Benefits Comprehensive health, dental, and vision coverage Flexible approach More ❯
Information Systems, or a related discipline. Desirable Experience Background or internship experience within financial services or technology. Exposure to Java. Experience managing on-premise or hybrid data infrastructure (e.g. Hadoop). Knowledge of workflow orchestration tools such as Apache Airflow. Postgraduate degree in Computer Science, Data Science, or related field. Benefits Comprehensive health, dental, and vision coverage Flexible approach More ❯
in data modelling, data warehousing, and ETL development. Hands-on experience with Azure Data Factory, Azure Data Lake, and Azure SQL Database. Exposure to big data technologies such as Hadoop, Spark, and Databricks. Experience with Azure Synapse Analytics or Cosmos DB. Familiarity with data governance frameworks (e.g., GDPR, HIPAA). Experience implementing CI/CD pipelines using Azure DevOps More ❯
in data modelling, data warehousing, and ETL development. Hands-on experience with Azure Data Factory, Azure Data Lake, and Azure SQL Database. Exposure to big data technologies such as Hadoop, Spark, and Databricks. Experience with Azure Synapse Analytics or Cosmos DB. Familiarity with data governance frameworks (e.g., GDPR, HIPAA). Experience implementing CI/CD pipelines using Azure DevOps More ❯
not all, but the majority of the below: Databases & SQL: SQL, Oracle DB, Postgres, SQL Server Messaging & Monitoring: ActiveMQ, Zabbix, Grafana, Ambari Cloud Platforms: AWS, Azure Big Data & Processing: Hadoop DevOps Tools: Jenkins, Puppet, BitBucket BPM & SOA: Oracle SOA, Oracle BPM, ActiviBPM Web & Application Servers: IIS Collaboration & Tracking: Jira, Confluence Other Technologies: CI tools and cloud-based technologies Desirable More ❯
predictive modelling, machine-learning, clustering and classification techniques, and algorithms Fluency in a programming language (Python, C,C++, Java, SQL) Familiarity with Big Data frameworks and visualization tools (Cassandra, Hadoop, Spark, Tableau More ❯
a data science team, mentoring junior colleagues and driving technical direction. Experience working with Agile methodologies in a collaborative team setting. Extensive experience with big data tools, such as Hadoop and Spark, for managing and processing large-scale datasets. Extensive experience with cloud platforms, particularly Microsoft Azure, for building and deploying data science solutions. Why Join? You'll be More ❯
e.g. MS SQL, Oracle) NoSQL technologies skills (e.g. MongoDB, InfluxDB, Neo4J) Data exchange and processing skills (e.g. ETL, ESB, API) Development (e.g. Python) skills Big data technologies knowledge (e.g. Hadoop stack) Knowledge in NLP (Natural Language Processing) Knowledge in OCR (Object Character Recognition) Knowledge in Generative AI (Artificial Intelligence) would be advantageous Experience in containerisation technologies (e.g. Docker) would More ❯