data processing) Apache Spark Streaming, Kafka or similar (for real-time data streaming) Experience using data tools in at least one cloud service - AWS, Azure or GCP (e.g. S3, EMR, Redshift, Glue, Azure Data Factory, Databricks, BigQuery, Dataflow, Dataproc) Would you like to join us as we work hard, have fun and make history? Apply for this job indicates More ❯
Proven experience as a Data Architect or Lead Data Engineer in AWS environments * Deep understanding of cloud-native data services: S3, Redshift, Glue, Athena, EMR, Kinesis, Lambda * Strong hands-on expertise in data modelling, distributed systems, and pipeline orchestration (Airflow, Step Functions) * Background in energy, trading, or financial markets is a strong plus * Excellent knowledge of Python, SQL, and More ❯
Proven experience as a Data Architect or Lead Data Engineer in AWS environments* Deep understanding of cloud-native data services: S3, Redshift, Glue, Athena, EMR, Kinesis, Lambda* Strong hands-on expertise in data modelling, distributed systems, and pipeline orchestration (Airflow, Step Functions)* Background in energy, trading, or financial markets is a strong plus* Excellent knowledge of Python, SQL, and More ❯
Leeds, West Yorkshire, England, United Kingdom Hybrid / WFH Options
Robert Walters
Key Skills & Experience: Proven experience as a Senior/Lead Data Engineer in a large-scale environment. Strong expertise with AWS data services (e.g., S3, Glue, Lambda, Redshift, Athena, EMR). Experience designing and building data lakes and modern data platforms. Proficiency with Python, SQL, and data pipeline orchestration tools (e.g., Airflow, dbt). Strong understanding of data modelling More ❯
to junior engineers Keep up-to-date with emerging data technologies and apply them where relevant The Skill Requirements: Hands-on experience with AWS services (Glue, Lambda, S3, Redshift, EMR) Strong skills in Python, SQL, PySpark and pipeline orchestration Proven understanding of data warehousing and data lakehouse concepts Excellent problem-solving skills with the ability to resolve performance bottlenecks More ❯
the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and code More ❯
Overview The Amazon Web Services Professional Services (ProServe) team is seeking a skilled Delivery Consultant to join our team at Amazon Web Services (AWS). In this role, you'll work closely with customers to design, implement, and manage AWS solutions that meet their technical requirements and business objectives. You'll be a key player in … and scalable AI solutions for business problems Interact with customers directly to understand the business problem, assist in the implementation of their ML ecosystem Leverage Foundation Models on Amazon Bedrock and Amazon SageMaker to meet performance needs Analyze large historical data to automate and optimize key processes Communicate clearly with attention to detail, translating rigorous mathematical … compelling customer proposals and present to executives; proficient English communication in technical and business settings Preferred Qualifications Experience with AWS services (Amazon SageMaker, Amazon Bedrock, EMR, S3, EC2); AWS Certification (Solutions Architect Associate, ML Engineer Associate) preferred Knowledge of AI/ML, generative AI; hands-on prompt engineering and deploying hosting Large Foundational Models Experience More ❯
Sterling, Virginia, United States Hybrid / WFH Options
Progression Inc
Python and Java 5+ years with Apache Spark (Spark SQL, Spark Streaming, data transformations, performance optimization, and tuning) 5+ years experience with cloud-based data platforms (Databricks or AWS EMR) Proficiency with Git and version control best practices Experience building and maintaining high-volume data pipelines Knowledge of data modeling and ETL best practices Familiarity with SQL and NoSQL More ❯
data pipelines and services • Ingest, transform, and model structured and unstructured data for analytics and ML • Work with technologies like Apache Spark, Apache Iceberg, Trino, NiFi, OpenSearch, and AWS EMR • Ensure data integrity, lineage, and security across the entire lifecycle • Collaborate with DevOps to deploy containerized data solutions using Kubernetes • Support Agile delivery, version control, and data governance activities More ❯
Maintaining the Apache Hadoop Ecosystem, especially utilizing HBase, MapReduce, and Spark. ETL processes utilizing Linux shell scripting, Perl, Python, and Apache Airflow. AWS services such as CloudWatch, CloudTrail, ELB, EMR, KMS, SQS, SNS, and Systems Manager. Experience in mobile platform development, such as web mobile, Android, microservices Working knowledge using proxies and proxy creation Vue.js, ASP.NET (C#), NodejsReact, JavaScript More ❯
and data integration concepts. Excellent communication skills to convey technical findings to non-technical stakeholders. Strong teamwork and collaboration abilities. Ability to translate business needs into technical specifications. EPIC EMR knowledge is a plus. Demonstrates good customer relations skills. Must be a team player. Skilled at problem solving, strong organization skills, strong attention to detail, and demonstrated ability to More ❯
disaster-recovery drills for stream and batch environments. Architecture & Automation Collaborate with data engineering and product teams to architect scalable, fault-tolerant pipelines using AWS services (e.g., Step Functions , EMR , Lambda , Redshift ) integrated with Apache Flink and Kafka . Troubleshoot & Maintain Python -based applications. Harden CI/CD for data jobs: implement automated testing of data schemas, versioned Flink More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Rackner
moving Agile DevSecOps team that builds secure, scalable data platforms-and get paid weekly. What You'll Do Build OpenAPI-compliant APIs, data schemas, & pipelines in AWS (S3, RDS, EMR) Develop with Python (FastAPI, Django, Flask) & JavaScript (Node.js, Vue, React) Deploy containerized workloads in Kubernetes (AWS EKS, Rancher) with CI/CD Apply DevSecOps + security-first practices from More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Rise Technical Recruitment Limited
Experience in a data-focused SRE, Data Platform, or DevOps role *Strong knowledge of Apache Flink, Kafka, and Python in production environments *Hands-on AWS experience with AWS (Lambda, EMR, Step Functions, Redshift, etc.) *Comfortable with monitoring tools, distributed systems debugging, and incident response Reference Number: BBBH259303 To apply for this role or for to be considered for further More ❯
strong problem-solving skills, and the ability to collaborate across business and technology teams. Key Responsibilities Design, develop, and maintain ETL/ELT data pipelines using AWS services (Glue, EMR, Redshift, S3, Athena, Lambda, Step Functions Develop and optimize large-scale data processing workflows using Python and PySpark. Collaborate with business stakeholders to translate data requirements into scalable technical … providing technical input for sprint planning and backlog grooming. Required Qualifications 10+ years of professional experience as a Data Engineer. Strong hands-on expertise in AWS Data Services: Glue, EMR, Redshift, Athena, S3, Lambda, Step Functions, CloudWatch. Proficiency in Python and PySpark for ETL, transformation, and automation. Strong experience in SQL and data modeling (star schema, snowflake schema, dimensional More ❯
software solutions at one or more layers of the technical stack (data, application, UI) Programming languages, particularly Python, Java, Javascript, TypeScript, SQL Cloud application and data deployment in AWS (EMR, EKS, ECR, RDS, etc.) DevOps tools and services (Kubernetes, Terraform, Docker, Packer, etc.) Integration with applications and data across a platform (e.g. APIs) Developing software within Agile methodologies Preferred More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Client Server
on data projects You have experience of establishing data analytics and supporting AI solutions You have a good working experience of AWS (e.g. S3, Kinesis, Glue, Redshift, Lambda and EMR) and/or Azure data services (e.g. ADF, Synapse, Fabric, Azure Functions) You have advanced client and stakeholder management skills What's in it for you: As a Data More ❯
processing engine such as DataBricks Workflows, Spark, Flink, Dask, dbt or similar - Experience building software and/or data pipelines in the AWS cloud (SageMaker Endpoints, ECS/EKS, EMR, Glue) Why Proofpoint Protecting people is at the heart of our award-winning lineup of cybersecurity solutions, and the people who work here are the key to our success. More ❯
strong problem-solving skills, and the ability to collaborate across business and technology teams. Key Responsibilities Design, develop, and maintain ETL/ELT data pipelines using AWS services (Glue, EMR, Redshift, S3, Athena, Lambda, Step Functions Develop and optimize large-scale data processing workflows using Python andPySpark. Collaborate with business stakeholders to translate data requirements into scalable technical solutions. … providing technical input for sprint planning and backlog grooming. Required Qualifications 10+ years of professional experience as a Data Engineer. Strong hands-on expertise in AWS Data Services: Glue, EMR, Redshift, Athena, S3, Lambda, Step Functions, CloudWatch. Proficiency in Python and PySpark for ETL, transformation, and automation. Strong experience in SQL and data modeling (star schema, snowflake schema, dimensional More ❯
as a technical liaison among system engineers, data scientists, analysts, and non-technical stakeholders to ensure aligned, mission-driven solutions. Key responsibilities include working with AWS cloud services (including EMR and Databricks), SQL database structures, and executing large-scale data migrations. The role also involves optimizing database architecture and performance, implementing DevSecOps practices, and building CI/CD pipelines More ❯
Maintaining the Apache Hadoop Ecosystem, especially utilizing HBase, MapReduce, and Spark. ETL processes utilizing Linux shell scripting, Perl, Python, and Apache Airflow. AWS services such as CloudWatch, CloudTrail, ELB, EMR, KMS, SQS, SNS, and Systems Manager. Vue.js, ASP.NET (C#), NodejsReact, JavaScript, HTML, CSS, PostgreSQL, Liquibase, Elasticsearch, and Git. Ansible Apache Niagra Files (NiFi) Apache TIKA Databricks and Lakehouse architecture More ❯
are seeking an experienced Data Engineer with expertise in AWS cloud technologies to design and build ETL pipelines , data warehouses , and data lakes . Key Skills: AWS services like EMR, Glue, Redshift, Kinesis, Lambda, DynamoDB (or equivalent open-source tools). Note: We require candidates who are eligible for SC Clearance or possess a higher level of clearance. More ❯
are seeking an experienced Data Engineer with expertise in AWS cloud technologies to design and build ETL pipelines , data warehouses , and data lakes . Key Skills: AWS services like EMR, Glue, Redshift, Kinesis, Lambda, DynamoDB (or equivalent open-source tools). Note: We require candidates who are eligible for SC Clearance or possess a higher level of clearance. More ❯
and mobile applications, work to maintain and extend the functionality of our web and mobile applications, as well as work on a variety of other technology projects such as EMR API integrations, AI features, and LMS offerings. This role requires an intimate understanding of what it takes to deliver and manage high-quality code in a complex, fast-paced More ❯