Cambridge, Cambridgeshire, United Kingdom Hybrid / WFH Options
OLI Systems
are building Galileo , a next-generation, cloud-native platform that will turn OLI's renowned engine into a global service. This is a ground-up re-architecture that adds distributedcomputing, AI-assisted simulations, and hardened infrastructure for always-on availability. Join a small, elite squad inside our Technology Team. You will work shoulder-to-shoulder with domain …/or TypeScript code. Advanced degree in computational chemistry, chemical engineering, or software engineering. Experience rewriting on-prem legacy engines into scalable cloud services. Contributions to open-source scientific-computing projects. Familiarity with distributedcomputing, scalable cloud services on cloud, or serverless architectures. Certifications in cloud security or AI/ML engineering. Key Competencies Technical Mastery & Architectural More ❯
technical skills and interpersonal competencies 20% - Provides direct global technical support for all requirements in developing strategy, planning, coordination, scripting, and implementation of releases, upgrades, or changes to our distributedcomputing environment. This includes participating in several small to medium-scale projects and/or tasks of various complexities across the enterprise. Your duties and responsibilities: Provide direct … technical support for all requirements in the management or changes to our Linux distributedcomputing Perform capacity planning, performance monitoring, tuning, and prediction. Provide high-level support for infrastructure with regards to the Linux operating systems, ensuring proper change management procedures and methodologies are followed. Some after-hours work will be required Position will require participation in an More ❯
that processes massive datasets, integrates complex geospatial data, and offers real-time or near-real-time insights. You will have the opportunity to work on cutting-edge technologies, including distributedcomputing, cloud infrastructure, and machine learning, to deliver a world-class platform that powers critical data-driven applications. Key Responsibilities: Platform Development & Optimization: Design, develop, and optimize the … or near-real-time data processing systems to deliver actionable insights to end-users. Optimize data flows and streaming analytics to ensure fast, low-latency decision-making capabilities. Cloud & Distributed Systems: Work with cloud platforms (AWS, GCP, Azure) to deploy and scale big data systems. Utilize containerization (e.g., Docker, Kubernetes) and cloud-native services to ensure flexible and scalable … engineering, with a focus on building and optimizing large-scale platforms for big data, data analytics, or geospatial data. Strong background in developing big data applications, data pipelines, and distributed systems. Proven experience working with geospatial data, including GIS, satellite imagery, and remote sensing data, and integrating it into data-driven applications. Familiarity with geospatial data formats (e.g., GeoJSON More ❯
agents in-product - to join their growing Data Engineering ream. Why You'll Love This Role: Tech at Scale : Work with hundreds of millions of data points daily, using distributed systems and advanced machine learning. Award-Winning Product : Recognised globally for innovation in AI. Remote-First Culture : Work from anywhere in the UK, with flexible hours and full autonomy … to deploy ML models. Use Gen AI tools to accelerate development and improve code quality. Contribute to the development of Gen AI agents in-product. Apply best practices in distributedcomputing, TDD, and system design. What We're Looking For: Strong experience with Python, Spark, Scala, and Java in a commercial setting. Solid understanding of distributed systems More ❯
Manchester, Lancashire, England, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
agents in-product - to join their growing Data Engineering ream. Why You'll Love This Role: Tech at Scale : Work with hundreds of millions of data points daily, using distributed systems and advanced machine learning. Award-Winning Product : Recognised globally for innovation in AI. Remote-First Culture : Work from anywhere in the UK, with flexible hours and full autonomy … to deploy ML models. Use Gen AI tools to accelerate development and improve code quality. Contribute to the development of Gen AI agents in-product. Apply best practices in distributedcomputing, TDD, and system design. What We're Looking For: Strong experience with Python, Spark, Scala, and Java in a commercial setting. Solid understanding of distributed systems More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
agents in-product - to join their growing Data Engineering ream. Why You'll Love This Role: Tech at Scale : Work with hundreds of millions of data points daily, using distributed systems and advanced machine learning. Award-Winning Product : Recognised globally for innovation in AI. Remote-First Culture : Work from anywhere in the UK, with flexible hours and full autonomy … to deploy ML models. Use Gen AI tools to accelerate development and improve code quality. Contribute to the development of Gen AI agents in-product. Apply best practices in distributedcomputing, TDD, and system design. What We're Looking For: Strong experience with Python, Spark, Scala, and Java in a commercial setting. Solid understanding of distributed systems More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
agents in-product - to join their growing Data Engineering ream. Why You'll Love This Role: Tech at Scale : Work with hundreds of millions of data points daily, using distributed systems and advanced machine learning. Award-Winning Product : Recognised globally for innovation in AI. Remote-First Culture : Work from anywhere in the UK, with flexible hours and full autonomy … to deploy ML models. Use Gen AI tools to accelerate development and improve code quality. Contribute to the development of Gen AI agents in-product. Apply best practices in distributedcomputing, TDD, and system design. What We're Looking For: Strong experience with Python, Spark, Scala, and Java in a commercial setting. Solid understanding of distributed systems More ❯
Birmingham, West Midlands, West Midlands (County), United Kingdom Hybrid / WFH Options
Tenth Revolution Group
agents in-product - to join their growing Data Engineering ream. Why You'll Love This Role: Tech at Scale : Work with hundreds of millions of data points daily, using distributed systems and advanced machine learning. Award-Winning Product : Recognised globally for innovation in AI. Remote-First Culture : Work from anywhere in the UK, with flexible hours and full autonomy … to deploy ML models. Use Gen AI tools to accelerate development and improve code quality. Contribute to the development of Gen AI agents in-product. Apply best practices in distributedcomputing, TDD, and system design. What We're Looking For: Strong experience with Python, Spark, Scala, and Java in a commercial setting. Solid understanding of distributed systems More ❯
or related field with relevant coursework in machine learning/statistics, software engineering principles, and database systems Core Skills: Python, ML frameworks (TensorFlow/PyTorch/scikit-learn), SQL, distributedcomputing, version control Experience: 1+ years in ML engineering or data engineering Mindset: Self-motivated with a growth mindset, adaptable to fast-paced startup environment, comfortable with ambiguity More ❯
or related field with relevant coursework in machine learning/statistics, software engineering principles, and database systems Core Skills: Python, ML frameworks (TensorFlow/PyTorch/scikit-learn), SQL, distributedcomputing, version control Experience: 1+ years in ML engineering or data engineering Mindset: Self-motivated with a growth mindset, adaptable to fast-paced startup environment, comfortable with ambiguity More ❯
hosting and managing our SDP microservice and connector infrastructure in AWS cloud. Design and implement big data technologies around Apache Hadoop, Kafka streaming, No SQL, Java/J2EE and distributedcomputing platforms. Participate in Agile development projects for enterprise-level systems component design and implementation. Apply enterprise software design for implementation of data services and middleware. You will More ❯
AI revolution! About DeepPCB: DeepPCB is InstaDeep's AI-powered Place & Route PCB (Printed Circuit Board) design tool. We use a combination of deep reinforcement learning and high-performance computing to automate and scale PCB place-and-route workflows, accelerating hardware innovation globally. We are looking for a Machine Learning Engineer to join the DeepPCB team and help push … engineers to bring ideas to life. Responsibilities: Develop scalable and efficient machine learning algorithms to tackle PCB place-and-route challenges. Adapt and optimize ML models for large-scale distributedcomputing environments (e.g., GPUs, multi-node clusters). Build, test, and deploy robust production-level ML systems integrated into the DeepPCB platform. Collaborate with research scientists, software engineers … thrive in a fast-paced, collaborative, and dynamic environment. Nice to haves: Prior experience with PCB design, EDA tools, or related optimization problems. Hands-on experience in high-performance computing environments (e.g., Kubernetes, Ray, Dask). Contributions to open-source projects, publications, or top placements in ML competitions (e.g., Kaggle). Expertise in related fields such as Computer Vision More ❯
skills required to triage and resolve complex production issues and operate well in a fast-paced, high-pressure environment. A propensity to automate manual tasks, appreciation for large-scale, distributedcomputing systems, and a willingness to develop using a wide range of languages and frameworks will be necessary to succeed in the role. As part of a global … to quickly identify scope and impact of issues during high-pressure situations Solid communication and interpersonal skills Ability to multi-task and prioritize tasks effectively Preferred Qualifications Experience with distributed systems design, maintenance, and troubleshooting. Hands-on experience with debugging and optimizing code, as well as automation. Knowledge of financial markets FIX protocol knowledge ABOUT GOLDMAN SACHS At Goldman More ❯
Face, LangChain, RAG pipelines). Strong programming skills in Python, SQL, and familiarity with cloud platforms (Azure, AWS, GCP). Expertise in data architecture, ETL/ELT pipelines, and distributedcomputing frameworks (e.g., Spark, Databricks). Excellent communication, stakeholder management, and team leadership skills. Nice to have (advantagous) Experience with MLOps, CI/CD for ML, and model More ❯
Face, LangChain, RAG pipelines). Strong programming skills in Python, SQL, and familiarity with cloud platforms (Azure, AWS, GCP). Expertise in data architecture, ETL/ELT pipelines, and distributedcomputing frameworks (e.g., Spark, Databricks). Excellent communication, stakeholder management, and team leadership skills. Nice to have (advantagous) Experience with MLOps, CI/CD for ML, and model More ❯
Azure. Software Engineering: Hands-on experience with foundational software engineering practices. Database Proficiency :Strong knowledge of SQL for querying and managing data. Scalability: Experience scaling computations using GPUs or distributedcomputing systems. ML Integration :Familiarity with exposing machine learning components through web services or wrappers (e.g., Flask in Python). Soft Skills: Strong communication and presentation skills to More ❯
ML-specific operators AI model serving experience with modern inference servers and API gateways for AI applications Nice to have: Infrastructure as Code experience with Terraform, Ansible, or CloudFormation Distributedcomputing experience with Databricks, Ray, or Spark for large-scale AI workloads AI safety & governance experience with model evaluation, bias detection, and responsible AI practices Multi-modal AI More ❯
expertise in Spark ML to work with a leading financial organisation on a global programme of work. The role involves predictive modeling, and deploying training and inference pipelines on distributed systems such as Hadoop. The ideal candidate will design, implement, and optimise machine learning solutions for large-scale data processing and predictive analytics. Role: Develop and implement machine learning … models using Spark ML for predictive analytics Design and optimise training and inference pipelines for distributed systems (e.g., Hadoop) Process and analyse large-scale datasets to extract meaningful insights and features Collaborate with data engineers to ensure seamless integration of ML workflows with data pipelines Evaluate model performance and fine-tune hyperparameters to improve accuracy and efficiency Implement scalable … solutions for real-time and batch inference Monitor and troubleshoot deployed models to ensure reliability and performance Stay updated with advancements in machine learning frameworks and distributedcomputing technologies Experience: Proficiency in Apache Spark and Spark MLlib for machine learning tasks Strong understanding of predictive modeling techniques (e.g., regression, classification, clustering) Experience with distributed systems like Hadoop More ❯
to design and build scalable, high-performance data solutions. Data Modelling & Warehouse Design : Proficiency in data modelling, warehouse design, and database optimization, with examples of logical and physical models. Distributed Data Systems : Experience in deploying, managing, and tuning distributed systems for optimal reliability and performance. Coding & Development Practices : Demonstrated coding expertise with modular, reusable, and efficient code in … engineering problems. Architecture for Scale : Design scalable, complex data architectures that provide cross-team value. Data Modelling & Governance : Establish standards in logical and physical data modelling and data governance. DistributedComputing : Employ parallel processing, streaming, and batch workflows to manage large data volumes effectively. ETL & Workflow Automation : Build ETL processes and automated workflows for efficient data movement. System More ❯
effectively across technical and non-technical teams Bonus Points For: Experience with Airflow, Airbyte, or other orchestration tools Familiarity with ingestion tools like Fivetran Experience working with Spark or distributedcomputing systems Exposure to AWS and broader cloud infrastructure Knowledge of reverse ETL workflows and tooling What Success Looks Like: You're delivering clean, well-modelled data quickly More ❯
through writing, visualisations, or presentations Strong organisational skills with experience in balancing multiple projects Familiarity with Posit Connect, workflow orchestration tools (e.g., Airflow), AWS services (e.g., SageMaker, Redshift), or distributedcomputing tools (e.g., Spark, Kafka) Experience in a media or newsroom environment Agile team experience Advanced degree in Maths, Statistics, or a related field What's in it More ❯
to expert in one or more technical areas. Design, implement and deliver performant and scalable algorithms based on state-of-the-art machine learning and neural network methodologies using distributedcomputing systems (CPUs, GPUs, TPUs, Cloud, etc.). Conduct rigorous data analysis and statistical modelling to explain and improve models. Report results clearly and efficiently, both internally and … on application. Nice to haves: Knowledge in areas around immunology, proteomics, and computer vision. Knowledge in molecular biology, biochemistry, structural biology, or a related discipline. Experience with high-performance computing or MLOps. Our commitment to our people We empower individuals to celebrate their uniqueness here at InstaDeep. Our team comes from all walks of life, and we're proud More ❯
Gloucester, Gloucestershire, South West, United Kingdom Hybrid / WFH Options
NSD
solutions and manage pipelines that transform diverse data sources into valuable insights for our client's National Security customers. You will collaborate with clients to solve complex challenges, utilising distributedcomputing techniques to handle large-scale, real-time, and unstructured data. Responsibilities include: Design and develop data pipelines, including ingestion, orchestration, and ETL processing (e.g., NiFi). Ensure More ❯
Gloucester, England, United Kingdom Hybrid / WFH Options
Searchability NS&D
solutions and manage pipelines that transform diverse data sources into valuable insights for our client’s National Security customers. You will collaborate with clients to solve complex challenges, utilising distributedcomputing techniques to handle large-scale, real-time, and unstructured data. Responsibilities include: Design and develop data pipelines, including ingestion, orchestration, and ETL processing (e.g., NiFi). Ensure More ❯
Experience in validating and QC'ing complex genomic datasets. Highly proficient in Python with solid command line knowledge and Unix skills. Highly proficient working with cloud environments (ideally Azure), distributedcomputing and optimising workflows and pipelines. Experience working with common data transformation and storage formats, e.g. Apache Parquet, Delta tables. Strong experience working with containerisation (e.g. Docker) and More ❯