working in the areas of ML and causal inference for downstream impact estimation. The ideal candidate will have knowledge of at least one of ray, spark or rapidsai framework to accelerate model training. A background in causal inference (e.g. Double ML) is a plus but not required. This is the More ❯
SOC 2). Collaborate with ML/AI Teams Package and deploy large‑language‑model (LLM) training jobs on distributed GPU clusters (Slurm, Ray, Kubeflow, or AWS SageMaker). Optimize model‑serving (Triton, vLLM, TorchServe) for low‑latency, high‑throughput inference. Cost & Performance Optimization Track cloud spend, right‑size resources More ❯
SOC 2). Collaborate with ML/AI Teams Package and deploy large‑language‑model (LLM) training jobs on distributed GPU clusters (Slurm, Ray, Kubeflow, or AWS SageMaker). Optimize model‑serving (Triton, vLLM, TorchServe) for low‑latency, high‑throughput inference. Cost & Performance Optimization Track cloud spend, right‑size resources More ❯
the pipeline. Collaborate with research to define data quality benchmarks . Optimize end-to-end performance across distributed data processing frameworks (e.g., Apache Spark, Ray, Airflow). Work with infrastructure teams to scale pipelines across thousands of GPUs . Work directly with the leadership on the data team roadmaps. Manage More ❯
to architectural decisions. What We’re Looking For: Strong Python programming skills (5+ years preferred). Deep experience with distributed systems (e.g., Kafka, Spark, Ray, Kubernetes). Hands-on work with big data technologies and architectures. Solid understanding of concurrency, fault tolerance, and data consistency. Comfortable in a fast-paced More ❯
london, south east england, United Kingdom Hybrid / WFH Options
Oliver Bernard
to architectural decisions. What We’re Looking For: Strong Python programming skills (5+ years preferred). Deep experience with distributed systems (e.g., Kafka, Spark, Ray, Kubernetes). Hands-on work with big data technologies and architectures. Solid understanding of concurrency, fault tolerance, and data consistency. Comfortable in a fast-paced More ❯
navigating hybrid infrastructure : some workloads will be on-prem, others cloud (large GPU clusters). Familiarity with distributed systems and container orchestration (e.g., Kubernetes, Ray). Experience working client-facing or in cross-functional teams — ideally within pharma/life sciences .1 A “get stuck in” attitude — this is a More ❯
navigating hybrid infrastructure : some workloads will be on-prem, others cloud (large GPU clusters). Familiarity with distributed systems and container orchestration (e.g., Kubernetes, Ray). Experience working client-facing or in cross-functional teams — ideally within pharma/life sciences .1 A “get stuck in” attitude — this is a More ❯
to some of the biggest names in the insurance industry. We are developing a modern real-time ML platform using technologies like Python, PyTorch, Ray, k8s (helm + flux), Terraform, Postgres and Flink on AWS. We are very big fans of Infrastructure-as-Code and enjoy Agile practices. As a More ❯
of those programs. The underlying models are a mix of causal and ML. The ideal candidate will have knowledge of at least one of ray, spark or rapidsai framework to accelerate model training. A background in causal inference (e.g. Double ML) is a plus but not required. This is the More ❯
Stack Our client is tech-agnostic and values adaptability. Current tools include: Backend : Python Frontend : TypeScript, React Infrastructure : Kubernetes, GCP Machine Learning : PyTorch, CUDA, Ray What’s on Offer Highly competitive base salary + commission + equity in a hyper-growth company 25 days holiday + public holidays Dynamic office More ❯
Stack Our client is tech-agnostic and values adaptability. Current tools include: Backend : Python Frontend : TypeScript, React Infrastructure : Kubernetes, GCP Machine Learning : PyTorch, CUDA, Ray What’s on Offer Highly competitive base salary + commission + equity in a hyper-growth company 25 days holiday + public holidays Dynamic office More ❯
breakdown of all the technologies we use: Backend: Python Frontend: Typescript and React Kubernetes for deployment GCP for underlying infrastructure Machine Learning: PyTorch, CUDA, Ray We encourage people from all backgrounds, cultures, and skill levels to apply. It is okay to not meet all requirements listed as we are looking More ❯
in Python and related ML frameworks such as JAX, Pytorch and XLA/MLIR. Experience with distributed training infrastructures (Kubernetes, Slurm) and associated frameworks (Ray). Experience using large-scale distributed training strategies. Hands on experience on training large model at scale and having contributed to the tooling and/ More ❯
Galytix (GX) is delivering on the promise of AI. GX has built specialised knowledge AI assistants for the banking and insurance industry. Our assistants are fed by sector-specific data and knowledge and easily adaptable through ontology layers to reflect More ❯
learning research and production. High-Performance Data Pipelines Develop and optimize distributed systems for data processing, including filtering, indexing, and retrieval, leveraging frameworks like Ray, Metaflow, Spark, or Hadoop. Synthetic Data Generation Build and orchestrate pipelines to generate synthetic data at scale, advancing research on cost-efficient inference and training … rendering engines, and/or other softwares. Distributed Computing & MLOps Demonstrated proficiency in setting up large-scale, robust data pipelines, using frameworks like Spark, Ray, or Metaflow. Comfortable with model versioning, and experiment tracking. Performance Optimization Good understanding of parallel and distributed computing. Experienced with setting up evaluation methods Cloud More ❯
Salt is proud to be partnered with a leading US-based technology company on the lookout for a talented Senior Software Developer to join their team. You'll be contributing to a large-scale analytics platform, building tailored SLA reporting More ❯