cost-optimal. Strong experience in different software testing paradigms to ensure consistency and correctness of our data. Bonus Skills Knowledge of real-time or stream processing systems (e.g. Kafka, Spark Streaming). Domain experience in energy, IoT, or working with unreliable/messy datasets. Frontend awareness—able to contribute to or reason about how data is visualized and interacted More ❯
cost-optimal. Strong experience in different software testing paradigms to ensure consistency and correctness of our data. Bonus Skills Knowledge of real-time or stream processing systems (e.g. Kafka, Spark Streaming). Domain experience in energy, IoT, or working with unreliable/messy datasets. Frontend awareness—able to contribute to or reason about how data is visualized and interacted More ❯
cost-optimal. Strong experience in different software testing paradigms to ensure consistency and correctness of our data. Bonus Skills Knowledge of real-time or stream processing systems (e.g. Kafka, Spark Streaming). Domain experience in energy, IoT, or working with unreliable/messy datasets. Frontend awareness—able to contribute to or reason about how data is visualized and interacted More ❯
Waltham On The Wolds, Leicestershire, United Kingdom
Mars, Incorporated and its Affiliates
constraints, context and risks. Solid knowledge of R or Python and related packages for data processing, visualisation and analysis. Awareness of techniques for querying and processing large data sets (Spark, SQL) and big data platforms (Azure). Understanding of coding best practices including clean, modular code, reproducibility, testing and version control (GitHub). Ability to effectively communicate complex data More ❯
University degree (IT/math) or equivalent experience The following additional qualifications are a significant plus: Kubernetes knowledge and operating experience Experience with big data stack components like Hadoop, Spark, Kafka, Nifi, Experience with data science/data analysis Knowledge of SRE/DevOP stacks - monitoring/system management tools (Prometheus, Ansible, ELK, ) Version control using git A day More ❯
App, AKS, Azure App Service, Web App. Good knowledge in real-time streaming applications preferably with experience in Kafka Real-time messaging or Azure Stream Analytics/Event Hub. Spark processing and performance tuning. File formats partitioning for e.g. Parquet, JSON, XML, CSV. Azure DevOps, GitHub actions. Hands-on experience in at least one of Python with knowledge of More ❯
history of successfully delivering large scale projects. Technologies we use (experience not required) AWS serverless architectures (Lambda, API Gateway, etc.) Kubernetes and container orchestration PostgreSQL and other relational databases Spark and distributed data processing Typescript and Java programming languages Infrastructure as Code (Terraform) Kafka and message queuing systems Git/GitHub, including GitHub Actions for CI/CD About More ❯
QUALIFICATIONS - 3+ years of experience in cloud architecture and implementation - Bachelor's degree in Computer Science, Engineering, related field, or equivalent experience - Experience in database (eg. SQL, NoSQL, Hadoop, Spark, Kafka, Kinesis) - Experience in consulting, design and implementation of serverless distributed solutions - Experience in software development with object oriented language PREFERRED QUALIFICATIONS - AWS experience preferred, with proficiency in a More ❯
monitoring tools such as New Relic or Grafana Understand the use of feature stores and related data technologies for operational machine learning products Are proficient with Python and have Spark knowledge. Have leadership experience either through previous management or mentorship. Have good communication skills. Nice to have Experience deploying LLMs and agent-based systems Our technology stack Python and More ❯
management and monitoring. Hands-on experience with AWS Have a good grasp of IaC (Infrastructure-as-code) tools like Terraform and CloudFormation. Previous exposure to additional technologies like Python, Spark, Docker, Kubernetes is desirable. Ability to develop across a diverse technology stack and willingness and ability to take on new technologies. Demonstrated experience participating on cross functional teams in More ❯
management and monitoring. Hands-on experience with AWS Have a good grasp of IaC (Infrastructure-as-code) tools like Terraform and CloudFormation. Previous exposure to additional technologies like Python, Spark, Docker, Kubernetes is desirable. Ability to develop across a diverse technology stack and willingness and ability to take on new technologies. Demonstrated experience participating on cross functional teams in More ❯
Power BI). Hands-on experience with SQL, Python, R, or similar languages for data analysis. Familiarity with cloud platforms (e.g., AWS, Google Cloud) and big data tools (e.g., Spark, Snowflake). Exceptional leadership, project management, and interpersonal skills with a proven ability to manage and scale teams. Strong business acumen with the ability to present complex data insights More ❯
do Passion for data and experience working within a data driven organization Hands-on experience with architecting, implementing, and performance tuning of: Data Lake technologies (e.g. Delta Lake, Parquet, Spark, Databricks) API & Microservices Message queues, streaming technologies, and event driven architecture NoSQL databases and query languages Data domain and event data models Data Modelling Logging and monitoring Container and More ❯
by a degree or interest in the legal domain. Ability to communicate with multiple stakeholders, including non-technical legal subject matter experts. Experience with big data technologies such as Spark, Hadoop, or similar. Experience conducting world-leading research, e.g. by contributions to publications at leading ML venues. Previous experience working on large-scale data processing systems. Strong software and More ❯
Swindon, Wiltshire, United Kingdom Hybrid / WFH Options
Zurich 56 Company Ltd
of Git. Experience using SQL and working with databases. Comfortable working with a variety of data sources, both structured and unstructured and very large datasets using distributed computing (e.g. spark). Experience working with cloud technology, ideally Microsoft Azure and/or AWS. Proven track record of development and deployment of machine learning algorithms, including supervised and unsupervised learning More ❯
Job Location COBALT PARK GO Job Description Are you ready to kickstart your tech career with the top-rated company for graduates in Fast Moving Consumer Goods (FMCG)? Do you want to be part of a company that develops the More ❯
Manchester Area, United Kingdom Hybrid / WFH Options
Opus Recruitment Solutions
Our client is a leading UK-based consultancy seeking a skilled professional to shape data strategies, mentor dynamic teams, and deliver cutting-edge solutions. With hands-on expertise in Spark, SQL, and cloud platforms like Azure, you’ll lead end-to-end projects, drive innovation, and collaborate with clients across industries. What You’ll Do: Lead complex data engineering … multi-project environments experience. Expertise in ETL, data modelling, and Azure Data Services. Experience in designing and implementing data pipelines, data lakes, and data warehouses. Hands-on experience with ApacheSpark and bonus points for Microsoft Fabric Any certifications are a bonus. Benefits: Competitive base salary Hybrid work once a week into their Central Manchester office 25 days More ❯
learning systems at scale You have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative teams Experience with TensorFlow … pyTorch, and/or other scalable Machine learning frameworks Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark Where You'll Be We offer you the flexibility to work where you work best! For this role, you can be within the EMEA region More ❯
/or teaching technical concepts to non-technical and technical audiences alike Passion for collaboration, life-long learning, and driving business value through ML Preferred Experience working with Databricks & ApacheSpark to process large-scale distributed datasets About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide - including Comcast, Condé Nast, Grammarly, and … Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. To learn more, follow Databricks on Twitter ,LinkedIn and Facebook . Benefits At Databricks, we strive to provide comprehensive benefits and perks that More ❯
ML) architectures on Azure, including: MLOps and model lifecycle management in Azure ML and Databricks. Model deployment, serving, and monitoring within a Lakehouse or Data Mesh framework. Proficiency in ApacheSpark, Python, SQL, and Scala for large-scale data processing. Excellent leadership, stakeholder engagement, and communication skills. Please apply if you have the above skill-set Carbon60, Lorien More ❯
code Experience working on distributed systems Strong knowledge of Kubernetes and Kafka Experience with Git, and Deployment Pipelines Having worked with at least one of the following stacks: Hadoop, ApacheSpark, Presto AWS Redshift, Azure Synapse or Google BigQuery Experience profiling performance issues in database systems Ability to learn and/or adapt quickly to complex issues Happy More ❯
science solutions in a commercial setting. MSc in Computer Science, Machine Learning, or a related field. Experience building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., ApacheSpark, Kafka, Airflow, dbt). Strong foundational knowledge of machine learning and deep learning algorithms, including deep neural networks, supervised/unsupervised learning, predictive analysis, and forecasting. Expert More ❯
science solutions in a commercial setting. MSc in Computer Science, Machine Learning, or a related field. Experience building data pipelines (realtime or batch) & data quality using modern toolchain (e.g., ApacheSpark, Kafka, Airflow, dbt). Strong foundational knowledge of machine learning and deep learning algorithms, including deep neural networks, supervised/unsupervised learning, predictive analysis, and forecasting. Expert More ❯
engineering workflows like CI/CD and containerised environments Skilled at working with both structured and unstructured data to unlock insights and power models Hands-on experience with Databricks, ApacheSpark, or similar tools used in large-scale data processing Exposure to machine learning model deployment using APIs or lightweight serving frameworks like Flask or Keras Familiarity with More ❯
engineering workflows like CI/CD and containerised environments Skilled at working with both structured and unstructured data to unlock insights and power models Hands-on experience with Databricks, ApacheSpark, or similar tools used in large-scale data processing Exposure to machine learning model deployment using APIs or lightweight serving frameworks like Flask or Keras Familiarity with More ❯