learning libraries in one or more programming languages. Keen interest in some of the following areas: Big Data Analytics (e.g. Google BigQuery/BigTable, ApacheSpark), Parallel Computing (e.g. ApacheSpark, Kubernetes, Databricks), Cloud Engineering (AWS, GCP, Azure), Spatial Query Optimisation, Data Storytelling with (Jupyter) Notebooks More ❯
learning libraries in one or more programming languages. Keen interest in some of the following areas: Big Data Analytics (e.g. Google BigQuery/BigTable, ApacheSpark), Parallel Computing (e.g. ApacheSpark, Kubernetes, Databricks), Cloud Engineering (AWS, GCP, Azure), Spatial Query Optimisation, Data Storytelling with (Jupyter) Notebooks More ❯
Working knowledge of two or more common Cloud ecosystems (AWS, Azure, GCP) with expertise in at least one. Deep experience with distributed computing with ApacheSpark and knowledge of Spark runtime internals. Familiarity with CI/CD for production deployments. Working knowledge of MLOps. Design and deployment … data, analytics, and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake, and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
Databricks Must Have Hands on experience on at least 2 Hyperscalers (GCP/AWS/Azure platforms) and specifically in Big Data processing services (ApacheSpark, Beam or equivalent). In-depth knowledge on key technologies like Big Query/Redshift/Synapse/Pub Sub/Kinesis … years’ experience in a similar role. Ability to lead and mentor the architects. Mandatory Skills [at least 2 Hyperscalers] GCP, AWS, Azure, Big data, Apachespark, beam on BigQuery/Redshift/Synapse, Pub Sub/Kinesis/MQ/Event Hubs, Kafka Dataflow/Airflow/ADF More ❯
Databricks Must Have Hands on experience on at least 2 Hyperscalers (GCP/AWS/Azure platforms) and specifically in Big Data processing services (ApacheSpark, Beam or equivalent). In-depth knowledge on key technologies like Big Query/Redshift/Synapse/Pub Sub/Kinesis … years’ experience in a similar role. Ability to lead and mentor the architects. Mandatory Skills [at least 2 Hyperscalers] GCP, AWS, Azure, Big data, Apachespark, beam on BigQuery/Redshift/Synapse, Pub Sub/Kinesis/MQ/Event Hubs, Kafka Dataflow/Airflow/ADF More ❯
to non-technical and technical audiences alike Passion for collaboration, life-long learning, and driving business value through ML Preferred Experience working with Databricks & ApacheSpark to process large-scale distributed datasets About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide - including … data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. To learn more, follow Databricks on Twitter ,LinkedIn and Facebook . Benefits At Databricks, we strive to provide More ❯
london, south east england, united kingdom Hybrid / WFH Options
ByteHire
ingestion, and processing pipelines pulling from sources such as social media and public data leaks. Supporting data processing and analysis at scale – think Elasticsearch, ApacheSpark (or equivalents), and PHP generators. Writing clean, testable code using PHPUnit, service containers, and SOLID principles. Participating in DevOps and deployment workflows … OpenAPI OAuth2, Laravel Sanctum Kafka, Redis AWS (ECS, ECR, RDS, CloudWatch) Terraform GitLab CI/CD Pipelines Docker SPA Frontend (React, Next.js, TailwindCSS) Elasticsearch, ApacheSpark Sentry, Grafana, AWS Logs Python, Java, Golang Requirements/experience: 5+ years of experience in backend PHP development, ideally with Laravel. Strong More ❯
south west london, south east england, united kingdom Hybrid / WFH Options
ByteHire
ingestion, and processing pipelines pulling from sources such as social media and public data leaks. Supporting data processing and analysis at scale – think Elasticsearch, ApacheSpark (or equivalents), and PHP generators. Writing clean, testable code using PHPUnit, service containers, and SOLID principles. Participating in DevOps and deployment workflows … OpenAPI OAuth2, Laravel Sanctum Kafka, Redis AWS (ECS, ECR, RDS, CloudWatch) Terraform GitLab CI/CD Pipelines Docker SPA Frontend (React, Next.js, TailwindCSS) Elasticsearch, ApacheSpark Sentry, Grafana, AWS Logs Python, Java, Golang Requirements/experience: 5+ years of experience in backend PHP development, ideally with Laravel. Strong More ❯
west london, south east england, united kingdom Hybrid / WFH Options
ByteHire
ingestion, and processing pipelines pulling from sources such as social media and public data leaks. Supporting data processing and analysis at scale – think Elasticsearch, ApacheSpark (or equivalents), and PHP generators. Writing clean, testable code using PHPUnit, service containers, and SOLID principles. Participating in DevOps and deployment workflows … OpenAPI OAuth2, Laravel Sanctum Kafka, Redis AWS (ECS, ECR, RDS, CloudWatch) Terraform GitLab CI/CD Pipelines Docker SPA Frontend (React, Next.js, TailwindCSS) Elasticsearch, ApacheSpark Sentry, Grafana, AWS Logs Python, Java, Golang Requirements/experience: 5+ years of experience in backend PHP development, ideally with Laravel. Strong More ❯
driving business value through ML Company first focus and collaborative individuals - we work better when we work together. Preferred Experience working with Databricks and ApacheSpark Preferred Experience working in a customer-facing role About Databricks Databricks is the data and AI company. More than 10,000 organizations … data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
AWS Certified Data Analytics - Specialty or AWS Certified Solutions Architect - Associate. Experience with Airflow for workflow orchestration. Exposure to big data frameworks such as ApacheSpark, Hadoop, or Presto. Hands-on experience with machine learning pipelines and AI/ML data engineering on AWS. Benefits: Competitive salary and More ❯
Kubeflow), and DevOps infrastructure (CI/CD pipelines, Terraform for infrastructure-as-code on cloud). Big Data & Analytics: Experience with big data technologies (Spark, Hadoop) or real-time data streaming (Kafka) in the context of analytics projects. This demonstrates ability to manage projects dealing with large-scale data More ❯
south west london, south east england, united kingdom
Inference Group
Kubeflow), and DevOps infrastructure (CI/CD pipelines, Terraform for infrastructure-as-code on cloud). Big Data & Analytics: Experience with big data technologies (Spark, Hadoop) or real-time data streaming (Kafka) in the context of analytics projects. This demonstrates ability to manage projects dealing with large-scale data More ❯
Kubeflow), and DevOps infrastructure (CI/CD pipelines, Terraform for infrastructure-as-code on cloud). Big Data & Analytics: Experience with big data technologies (Spark, Hadoop) or real-time data streaming (Kafka) in the context of analytics projects. This demonstrates ability to manage projects dealing with large-scale data More ❯
with TensorFlow, PyTorch, Scikit-learn, etc. is a strong plus. You have some experience with large scale, distributed data processing frameworks/tools like Apache Beam, ApacheSpark, or even our open source API for it - Scio, and cloud platforms like GCP or AWS. You care about More ❯
architectures for ML frameworks in complex problem spaces in collaboration with product teams Experience with large scale, distributed data processing frameworks/tools like Apache Beam, ApacheSpark, and cloud platforms like GCP or AWS Where You'll Be We offer you the flexibility to work where More ❯
delivery across a range of projects, including data analysis, extraction, transformation, and loading, data intelligence, data security and proven experience in their technologies (e.g. Spark, cloud-based ETL services, Python, Kafka, SQL, Airflow) You have experience in assessing the relevant data quality issues based on data sources & uses cases More ❯
team-oriented environment. Preferred Skills: Experience with programming languages such as Python or R for data analysis. Knowledge of big data technologies (e.g., Hadoop, Spark) and data warehousing concepts. Familiarity with cloud data platforms (e.g., Azure, AWS, Google Cloud) is a plus. Certification in BI tools, SQL, or related More ❯
Purview, or Informatica, including projects around lineage, cataloging, and quality rules. Strong hands-on development experience in SQL and Python, with working knowledge of Spark or other distributed data processing frameworks. Design, development and implementation of distributed data solutions using API and microservice-based architecture. Deep understanding of ETL More ❯
Purview, or Informatica, including projects around lineage, cataloging, and quality rules. Strong hands-on development experience in SQL and Python, with working knowledge of Spark or other distributed data processing frameworks. Design, development and implementation of distributed data solutions using API and microservice-based architecture. Deep understanding of ETL More ❯
Purview, or Informatica, including projects around lineage, cataloging, and quality rules. Strong hands-on development experience in SQL and Python, with working knowledge of Spark or other distributed data processing frameworks. Design, development and implementation of distributed data solutions using API and microservice-based architecture. Deep understanding of ETL More ❯
engineering, including infrastructure-as-code (e.g., Terraform, CloudFormation), CI/CD pipelines, and monitoring (e.g., CloudWatch, Datadog). Familiarity with big data technologies like ApacheSpark, Hadoop, or similar. ETL/ELT tools and creating common data sets across on-prem (IBMDatastage ETL) and cloud data stores Leadership More ❯
or a related field, with a focus on building scalable data systems and platforms. Strong expertise with modern data tools and frameworks such as Spark , dbt , Airflow OR Kafka , Databricks , and cloud-native services (AWS, GCP, or Azure). Deep understanding of data modeling , distributed systems , streaming architectures , and More ❯
programming languages such as Python or Java. Understanding of data warehousing concepts and data modeling techniques. Experience working with big data technologies (e.g., Hadoop, Spark) is an advantage. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Benefits Enhanced leave - 38 days inclusive of 8 UK Public More ❯
programming languages such as Python or Java Understanding of data warehousing concepts and data modeling techniques Experience working with big data technologies (e.g., Hadoop, Spark) is an advantage Excellent problem-solving and analytical skills Strong communication and collaboration skills Benefits Enhanced leave - 38 days inclusive of 8 UK Public More ❯