Synapse Analytics with Spark and SQL, Azure functions with Python, Azure Purview, and Cosmos DB. They are also proficient in Azure Event Hub and Streaming Analytics, Managed Streaming for Apache Kafka, Azure DataBricks with Spark, and other open source technologies like Apache Airflow and dbt, Spark/Python, or Spark/Scala. Preferred Education Bachelor's Degree Required More ❯
and maintain Data Pipelines using Python. Develop real-time streaming features using big data tools such as Spark. SKILLS AND EXPERIENCE Extensive experience using big data tools such as Apache Spark. Experience working in and maintaining an AWS database. Strong Python coding background. Good knowledge of working with SQL. THE BENEFITS Generous Holiday plan. Career development plan. Flexible working. More ❯
data quality improvement initiatives and proactively resolve data inconsistencies. · Participate in troubleshooting and performance tuning of ETL jobs and workflows. Required Skills & Qualifications: · Proven experience with Talend, Python, and Apache Spark. · Strong understanding of relational databases and Big Data ecosystems (Hive, Impala, HDFS). · Solid experience in data warehousing and data modelling techniques. · Familiarity with data quality management and More ❯
Newcastle Upon Tyne, England, United Kingdom Hybrid / WFH Options
In Technology Group
warehousing. Proficiency in Python or another programming language used for data engineering. Experience with cloud platforms (e.g., Azure, AWS, or GCP) is highly desirable. Familiarity with tools such as Apache Airflow, Spark, or similar is a plus. What’s On Offer: Competitive salary between £45,000 – £55,000 , depending on experience. Flexible hybrid working – 3 days on-site in More ❯
and familiarity with templating approaches (e.g., Jinja). Hands-on experience with cloud technologies, ideally within AWS environments. Proven ability to work with orchestration platforms-experience with tools like Apache Airflow is a plus. Comfortable developing CI/CD workflows, ideally using tools such as GitHub Actions. Experience building and maintaining modern data pipelines and infrastructure. Cooperative approach to More ❯
Senior Software engineer to join their team in London on a full-time basis What You’ll Do Architect and implement high-performance data processing systems in Rust Leverage Apache Arrow and Parquet for in-memory and on-disk data efficiency Integrate and extend systems like DataFusion, ClickHouse, and DuckDB Design low-latency pipelines for analytical workloads Collaborate with More ❯
Senior Software engineer to join their team in London on a full-time basis What You’ll Do Architect and implement high-performance data processing systems in Rust Leverage Apache Arrow and Parquet for in-memory and on-disk data efficiency Integrate and extend systems like DataFusion, ClickHouse, and DuckDB Design low-latency pipelines for analytical workloads Collaborate with More ❯
Senior Software engineer to join their team in London on a full-time basis What You’ll Do Architect and implement high-performance data processing systems in Rust Leverage Apache Arrow and Parquet for in-memory and on-disk data efficiency Integrate and extend systems like DataFusion, ClickHouse, and DuckDB Design low-latency pipelines for analytical workloads Collaborate with More ❯
London, England, United Kingdom Hybrid / WFH Options
KennedyPearce Consulting
S3, Redshift, RDS, Glue, Lambda, IAM). Strong expertise in Terraform Proficient in SQL for querying relational databases and handling large datasets. Experience with data pipeline orchestration tools (e.g., Apache Airflow, AWS Step Functions). Familiarity with CI/CD pipelines and version control systems (e.g., Git). Knowledge of data warehousing concepts and best practices. Benefits : Competitive salary More ❯
Senior Software engineer to join their team in London on a full-time basis What You’ll Do Architect and implement high-performance data processing systems in Rust Leverage Apache Arrow and Parquet for in-memory and on-disk data efficiency Integrate and extend systems like DataFusion, ClickHouse, and DuckDB Design low-latency pipelines for analytical workloads Collaborate with More ❯
level of proficiency in Python and SQL A deep understanding of generative AI and experience building applications with large foundation models Proficiency in Databricks, BitBucket/Jira/Confluence, Apache Spark (PySpark) and AWS is advantageous Degree/Master’s educated in Data Science or a related subject – e.g. Statistics, Computer Science, etc. A working knowledge of fundamental equity More ❯
pipelines, and implementing scalable solutions that meet the evolving needs of the business. Utilise your strong background in data engineering, combined with your existing experience using SQL, Python and Apache Spark in production environments. The role will entail strong problem-solving skills, attention to detail, and the ability to work independently while collaborating closely with internal and external stakeholders. More ❯
code Experience working on distributed systems Strong knowledge of Kubernetes and Kafka Experience with Git, and Deployment Pipelines Having worked with at least one of the following stacks: Hadoop, Apache Spark, Presto AWS Redshift, Azure Synapse or Google BigQuery Experience profiling performance issues in database systems Ability to learn and/or adapt quickly to complex issues Happy to More ❯
code Desired Skills (Bonus Points): Proven experience in recommender systems, behavioural AI, and/or reinforcement learning. Building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., Apache Spark, Kafka, Airflow, dbt). PhD in Computer Science, Machine Learning, or a closely related field What We Offer: Opportunity to build technology that will transform millions of shopping More ❯
code Desired Skills (Bonus Points): Proven experience in recommender systems, behavioural AI, and/or reinforcement learning. Building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., Apache Spark, Kafka, Airflow, dbt). PhD in Computer Science, Machine Learning, or a closely related field What We Offer: Opportunity to build technology that will transform millions of shopping More ❯
Moderate experience in virtualization technologies like VMWare, OpenStack, Xen Knowledge of containerization (basic level) Exposure to cloud platforms such as AWS, GCP, Azure, or OPC Adept at Kafka, Cassandra, Apache Spark, HDFS Monitoring experience with Prometheus, Nagios, Icinga Experience with logging tools like Splunk or ELK stack Handling configuration management with Ansible, Terraform, Chef, or Puppet Comprehensive experience in More ❯
code Experience working on distributed systems Strong knowledge of Kubernetes and Kafka Experience with Git, and Deployment Pipelines Having worked with at least one of the following stacks: Hadoop, Apache Spark, Presto Experience profiling performance issues in database systems Ability to learn and/or adapt quickly to complex issues Happy to collaborate with a wide group of stakeholders More ❯
with DevOps practices for data engineering, including infrastructure-as-code (e.g., Terraform, CloudFormation), CI/CD pipelines, and monitoring (e.g., CloudWatch, Datadog). Familiarity with big data technologies like Apache Spark, Hadoop, or similar. ETL/ELT tools and creating common data sets across on-prem (IBMDatastage ETL) and cloud data stores Leadership & Strategy: Lead Data Engineering team(s More ❯
science solutions in a commercial setting. MSc in Computer Science, Machine Learning, or a related field. Experience building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., Apache Spark, Kafka, Airflow, dbt). Strong foundational knowledge of machine learning and deep learning algorithms, including deep neural networks, supervised/unsupervised learning, predictive analysis, and forecasting. Expert-level More ❯
science solutions in a commercial setting. MSc in Computer Science, Machine Learning, or a related field. Experience building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., Apache Spark, Kafka, Airflow, dbt). Strong foundational knowledge of machine learning and deep learning algorithms, including deep neural networks, supervised/unsupervised learning, predictive analysis, and forecasting. Expert-level More ❯
Engineering (open to professionals from various data eng. backgrounds — data pipelines, ML Eng, data warehousing, analytics engineering, big data, cloud etc.) Technical Exposure: Experience with tools like SQL, Python, Apache Spark, Kafka, Cloud platforms (AWS/GCP/Azure), and modern data stack technologies Formal or Informal Coaching Experience: Any previous coaching, mentoring, or training experience — formal or informal More ❯
and quality rules What Makes a Great Candidate: Experience in a data specialist role, a passion for working with data and helping stakeholders Highly proficient in SQL, Python, and Apache Spark, with demonstrable work experience using these tools in a production context A minimum 2.1 degree obtained in Computer Science, or a related field, ideally from a Russell Group More ❯
London, England, United Kingdom Hybrid / WFH Options
FIND | Creating Futures
Engineering (open to professionals from various data eng. backgrounds — data pipelines, ML Eng, data warehousing, analytics engineering, big data, cloud etc.) Technical Exposure: Experience with tools like SQL, Python, Apache Spark, Kafka, Cloud platforms (AWS/GCP/Azure), and modern data stack technologies Formal or Informal Coaching Experience: Any previous coaching, mentoring, or training experience — formal or informal More ❯
Edinburgh, Scotland, United Kingdom Hybrid / WFH Options
M-KOPA
engineering, for machine learning or general analytics use case. Additionally, having experience with Kubernetes or other platforms for containerized applications as well as working with orchestration systems such as Apache Airflow would be essential to succeed in this role. The ideal candidate for this role would need to have proficiency in programming languages (Python, C#, Java, etc.) as well More ❯
on real client use cases. Proficient in one of the deep learning stacks such as PyTorch or Tensorflow. Working knowledge of parallelisation and async paradigms in Python, Spark, Dask, Apache Ray. An awareness and interest in economic, financial and general business concepts and terminology. Excellent written and verbal command of English. Strong problem-solving, analytical and quantitative skills. A More ❯