and stream processing. Hands-on experience building and operation for data lake using one or more of the following big data frameworks or services: Spark, Kafka, Airflow, DBT, Debezium, AWS Athena, AWS Glue, Delta lake/Iceberg etc. Experience with Kubernetes, Docker, Terraform or other cluster management solutions on More ❯
in the legal domain. Ability to communicate with multiple stakeholders, including non-technical legal subject matter experts. Experience with big data technologies such as Spark, Hadoop, or similar. Experience conducting world-leading research, e.g. by contributions to publications at leading ML venues. Previous experience working on large-scale data More ❯
in the legal domain. Ability to communicate with multiple stakeholders, including non-technical legal subject matter experts. Experience with big data technologies such as Spark, Hadoop, or similar. Experience conducting world-leading research, e.g. by contributions to publications at leading ML venues. Previous experience working on large-scale data More ❯
3rd party packages, including MLFlow, Seldon for ML model tracking and deployment, Kubernetes for hosting models, Argo and Git for CI/CD automation, Spark for big data processing. This is a rapidly changing field and we are deeply involved in open source community to help shape the technology … to adopt in order to enhance our platform. What you'll do: Develop across our evolving technology stack - we're using Python, Java, Kubernetes, ApacheSpark, Postgres, ArgoCD, Argo Workflow, Seldon, MLFlow and more. We are migrating into AWS cloud and adopting many services that are available in … desire to write clean, correct and efficient code. Sense of ownership, urgency and pride in your work. Experience with Python, Java, Docker, Kubernetes, Argo, Spark and AWS cloud services a plus. Exposure to Machine Learning practices a plus. We strive to create an accessible and inclusive experience for all More ❯
Purview or equivalent for data governance and lineage tracking Experience with data integration, MDM, governance, and data quality tools . Hands-on experience with ApacheSpark, Python, SQL, and Scala for data processing. Strong understanding of Azure networking, security, and IAM , including Azure Private Link, VNETs, Managed Identities More ❯
Purview or equivalent for data governance and lineage tracking Experience with data integration, MDM, governance, and data quality tools . Hands-on experience with ApacheSpark, Python, SQL, and Scala for data processing. Strong understanding of Azure networking, security, and IAM , including Azure Private Link, VNETs, Managed Identities More ❯
Nottingham, Nottinghamshire, United Kingdom Hybrid / WFH Options
Experian Group
Significant experience of programming using Scala and Python Experience of using Terraform to provision and deploy cloud services and components Experience of developing on ApacheSpark Experience of developing with AWS cloud services including (but not limited to) AWS Glue, S3, Step Functions, Lambdas, EventBridge and SQS BDD More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
the following certifications are desired: Certified Cloud Security Professional (CCSP) , GIAC Security Essentials Certification (GSEC), or CompTIA Cybersecurity Analyst (CySA+) Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
networking, storage, to applications Design and deploy clouds and applications for external customers Architect cloud infrastructure solutions like Kubernetes, Kubeflow, OpenStack, Ceph, Hadoop, and Spark, either on-premises or in public cloud platforms (AWS, Azure, Google Cloud) Foster a collaborative engineering culture aligned with company values Work remotely from More ❯
Familiarity with cloud platforms like AWS, GCP, or Azure. Strong written and spoken English skills. Bonus Experience: Experience with big data tools (e.g., Hadoop, Spark) and distributed computing. Knowledge of NLP techniques and libraries. Familiarity with Docker, Kubernetes, and deploying machine learning models in production. Experience with visualization tools More ❯
Columbia, Maryland, United States Hybrid / WFH Options
SilverEdge
days per week. Flexibility is essential to adapt to schedule changes if needed. Desired Qualifications Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. About SilverEdge More ❯
clients to deliver these analytical solutions Collaborate with stakeholders and customers to ensure successful project delivery Write production-ready code in SQL, Python, and Spark following software engineering best practices Coach team members in machine learning and statistical modelling techniques Who we are looking for We are looking for More ❯
and understanding of a broad range of data-oriented technologies and their applicability. Experience with one or more data warehouse platforms (preferably Databricks and Spark). Understanding of the principles, practices and techniques of Data Engineering and Machine Learning. Working knowledge of one or more data-oriented programming languages More ❯
MONTH INITIAL CONTRACT Seeking a Data Engineer who has a strong understanding of data concepts - data types, data structures, schemas (both JSON and Spark), schema management etc - Strong understanding of complex JSON manipulation - Experience working with Data Pipelines using a custom Python/PySpark frameworks - Strong understanding of the … Lake - Bash (both CLI usage and scripting) - Git - Markdown - Scala (bonus, not compulsory) - Azure SQL Server as a HIVE Metastore (bonus) Technologies - Azure Databricks - ApacheSpark - Delta Tables - Data processing with Python - PowerBI (Integration/Data Ingestion) - JIRA Due to the nature and urgency of this post, candidates More ❯
Leeds, West Yorkshire, Yorkshire, United Kingdom Hybrid / WFH Options
Evri
flexibility for remote work. Office visits required for key meetings and collaboration sessions. Key Responsibilities: Develop and maintain scalable data pipelines using Databricks and ApacheSpark to process logistics and delivery data Design ETL workflows that integrate data from multiple delivery and warehouse systems Development of Data Marts More ❯
Saffron Walden, Essex, South East, United Kingdom Hybrid / WFH Options
EMBL-EBI
Experience in developing web infrastructure (Solr, kubernetes) Experience in git and basic Unix Commands You may also have Experience with large data processing technologies (ApacheSpark) Other helpful information: The team work in a hybrid working pattern and spends 2 days per week in office Apply now! Benefits More ❯
Birmingham, Staffordshire, United Kingdom Hybrid / WFH Options
Investigo
advanced visualisations, ML model interpretation, and KPI tracking. Deep knowledge of feature engineering, model deployment, and MLOps best practices. Experience with big data processing (Spark, Hadoop) and cloud-based data science environments. Other: Ability to integrate ML workflows into large-scale data pipelines. Strong experience in data preprocessing, feature More ❯
Birmingham, West Midlands, West Midlands (County), United Kingdom Hybrid / WFH Options
Investigo
advanced visualisations, ML model interpretation, and KPI tracking. Deep knowledge of feature engineering, model deployment, and MLOps best practices. Experience with big data processing (Spark, Hadoop) and cloud-based data science environments. Other: Ability to integrate ML workflows into large-scale data pipelines. Strong experience in data preprocessing, feature More ❯
causal inference, etc., is beneficial Programming experience with Python, Scala, Java, or strong knowledge of another object-oriented or functional language Experience with AWS, Spark, SQL Market-leading base salary and restricted stock units Market-leading benefits package Hybrid work environment (London office and remote work) #J-18808-Ljbffr More ❯
services (SageMaker, Lambda, Glue, Redshift, Athena, S3). Strong understanding of ML model development, deployment, and monitoring . Experience with big data processing frameworks (Spark, EMR, Kafka) is a plus. Background in FinTech or financial services is advantageous. What’s on Offer? Competitive salary of More ❯
seamless data flow and processing. Use your expertise in SQL , along with experience in relational and NoSQL databases. Transform and process data using Python , Spark , or Scala . Implement secure, scalable ETL frameworks that deliver value and insights. What's In It For You? Competitive salary. Flexible Hybrid Working More ❯
. Experience in using cloud-native services for data engineering and analytics. Experience with distributed systems, serverless data pipelines, and big data technologies (e.g., Spark, Kafka). Ability to define and enforce data governance standards. Experience in providing architectural guidance, mentorship and leading cross-functional discussions to align on More ❯
london, south east england, united kingdom Hybrid / WFH Options
Formula Recruitment
. Experience in using cloud-native services for data engineering and analytics. Experience with distributed systems, serverless data pipelines, and big data technologies (e.g., Spark, Kafka). Ability to define and enforce data governance standards. Experience in providing architectural guidance, mentorship and leading cross-functional discussions to align on More ❯
working with Federal Government and/or DoD clients. • Prior experience with NoSQL technologies. • Proficient in containerization systems (Kubernetes). • Exposure to Python, JupyterLab, Spark, and/or SQL for developing ETL applications. • Confident with at least one of the major Cloud providers (AWS). • Comfortable with Unix/ More ❯
Manchester, Lancashire, United Kingdom Hybrid / WFH Options
iO Associates
Skills Familiarity with cloud platforms such as Azure (e.g., Azure ML, Data Factory). Experience in big data environments and distributed computing frameworks (e.g., Spark). Knowledge of business intelligence tools and their integration with data science workflows. Prior experience mentoring or leading a team of data scientists. Why More ❯