Engineer using cloud platform technologies, alongside experience with a variety of database technologies including Oracle, Postgres and MSSQLServer; Strong expertise in AWS services including AWS DMS, S3, Lambda, Glue, EMR, Redshift, and IAM. Proficient in Apache Spark (batch and/or streaming) and big data processing. Solid experience with SQL and performance tuning in data warehouse environments. Hands-on … experience with Amazon Redshift or equivalent, including table design, workload management, and implementing Redshift Spectrum. Experience building ETL/ELT pipelines using tools like AWS Glue, EMR, or custom frameworks. Familiarity with data modeling concepts. Excellent problem-solving and communication skills. Proficiency in Java and data pipeline development. Familiarity with version control systems (e.g., Git) and agile More ❯
and Kinesis. Work with structured and unstructured data from multiple sources, ensuring efficient data ingestion, transformation, and storage. Develop and optimize data lake and data warehouse solutions using Amazon S3, Redshift, Athena, and Lake Formation. Implement data governance, security, and compliance best practices, including IAM roles, encryption, and access controls. Monitor and optimize performance of data workflows using … a strong focus on AWS cloud technologies. Proficiency in Python, PySpark, SQL, and AWS Glue for ETL development. Hands-on experience with AWS data services, including Redshift, Athena, Glue, EMR, and Kinesis. Strong knowledge of data modeling, warehousing, and schema design. Experience with event-driven architectures, streaming data, and real-time processing using Kafka or Kinesis. Expertise in IaC More ❯
Full-time | Hybrid | London | Up to £95,000 + Bonus & Excellent Benefits | Key Responsibilities: Design, develop, and implement advanced data pipelines and ETL/ELT workflows using cloud-native services such as AWS Glue, Lambda, S3, Redshift, and EMR. Act More ❯
Drive the design development and implementation of complex data pipelines and ETL/ELT processes using cloud-native technologies (e.g. AWS Glue AWS Lambda AWS S3 AWS Redshift AWS EMR). Develop and maintain data quality checks data validation rules and data lineage documentation. Collaborate with data analysts data scientists business stakeholders and product owners to understand their data More ❯
role with advanced experience in at least most of the following technologies Cloud Platforms: AWS (Amazon Web Services): Knowledge of services like S3, EC2, Lambda, RDS, Redshift, EMR, SageMaker, Glue, and Kinesis. Azure: Proficiency in services like Azure Blob Storage, Azure Data Lake, VMs, Azure Functions, Azure SQL Database, HDInsight, and Azure Machine Learning Studio. Data Storage … Azure Purview. Data Security Tools: AWS Key Management Service (KMS), Azure Key Vault. Data Analytics & BI: Visualization Tools: Tableau, Power BI, Looker, and Grafana. Analytics Services: AWS Athena, Amazon QuickSight, Azure Stream Analytics. Development & Collaboration Tools: Version Control: Git (and platforms like GitHub, GitLab). CI/CD Tools: Jenkins, Travis CI, AWS CodePipeline, Azure DevOps. Other Key More ❯
data processing) Apache Spark Streaming, Kafka or similar (for real-time data streaming) Experience using data tools in at least one cloud service - AWS, Azure or GCP (e.g. S3, EMR, Redshift, Glue, Azure Data Factory, Databricks, BigQuery, Dataflow, Dataproc Would you like to join us as we work hard, have fun and make history? Apply for this job indicates More ❯
data processing) Apache Spark Streaming, Kafka or similar (for real-time data streaming) Experience using data tools in at least one cloud service - AWS, Azure or GCP (e.g. S3, EMR, Redshift, Glue, Azure Data Factory, Databricks, BigQuery, Dataflow, Dataproc) Would you like to join us as we work hard, have fun and make history? Apply for this job indicates More ❯
of relevant industry experience in a data engineering capacity Experience translating business needs to scalable data solutions. Experience building highly scalable data pipelines (batch and streaming) using AirFlow, Spark, EMR, Kafka, AWS Kinesis. Experience designing and developing solutions on AWS, including infrastructure as Code (e.g. Cloud Formation, Terraform, AWS CDK) Experience with AWS compute resources such as Lambda and More ❯
complex business requirements and drive decision-making. Your skills and experience Proficiency with AWS Tools: Demonstrable experience using AWS Glue, AWS Lambda, Amazon Kinesis, AmazonEMR , Amazon Athena, Amazon DynamoDB, Amazon Cloudwatch, Amazon SNS and AWS Step Functions. Programming Skills: Strong experience with modern programming languages such More ❯
London, England, United Kingdom Hybrid / WFH Options
RED Global
the following: Proven experience as an AWS Data SME or AWS Data Engineer , working extensively with AWS cloud services. Expertise in AWS Redshift, Glue, Lambda, Terraform, Kinesis, Athena, and EMR . Strong ETL/ELT development and data warehousing experience. Proficiency in Python, Java, or Scala for data processing and automation. In-depth knowledge of SQL, Apache Kafka, and … Amazon RDS . Experience in data security, governance, and compliance best practices . Familiarity with CI/CD pipelines, DevOps methodologies, and monitoring/logging best practices . Strong problem-solving skills , with the ability to work in a collaborative and fast-paced environment . Preferred Qualifications: AWS Certified Data Analytics - Specialty or AWS Certified Solutions Architect/ More ❯
be working on complex data problems in a challenging and fun environment, using some of the latest Big Data open-source technologies like Apache Spark, as well as Amazon Web Service technologies including ElasticMapReduce, Athena and Lambda to develop scalable data solutions. Key Responsibilities: Adhering to Company Policies and Procedures with respect to Security, Quality and Health More ❯
ETL) data from various sources into AWS data stores for both batch and streaming data ingestion. AWS Expertise: Utilize your expertise in AWS services such as AmazonEMR, S3, AWS Glue, Amazon Redshift, AWS Lambda, and more to build and optimize data solutions. Data Modeling: Design and implement data models to support analytical and reporting … to understand data requirements and deliver high-quality data solutions. Automation: Implement automation processes and best practices to streamline data workflows and reduce manual interventions. Must have: AWS, ETL, EMR, GLUE, Spark/Scala, Java, Python. Good to have: Cloudera – Spark, Hive, Impala, HDFS, Informatica PowerCenter, Informatica DQ/DG, Snowflake Erwin. Qualifications: Bachelor's or Master's degree More ❯
Milton Keynes, England, United Kingdom Hybrid / WFH Options
Santander
to be successful in this role: Experience developing, testing, and deploying data pipelines, data lakes, data warehouses, and data marts using ideally AWS services such as S3, Glue, Athena, EMR, Kinesis, and Lambda Understanding of the principles behind designing and implementing data lake, lake house and/or data mesh architecture Problem solving skill with basic knowledge of the … with team members, stakeholders and end users conveying technical concepts in a comprehensible manner Skills across the following data competencies: SQL (AWS Athena/Hive/Snowflake) Hadoop/EMR/Spark/Scala Data structures (tables, views, stored procedures) Data Modelling - star/snowflake Schemas, efficient storage, normalisation Data Transformation DevOps - data pipelines Controls - selection and build Reference More ❯
alongside Associates, Mid-level Engineers, and Senior/Principal engineers. What You'll Do Design, build, and scale data pipelines and services using AWS Glue, Lambda, Step Functions, S3, EMR, Athena, and more. Lead projects involving serverless, event-driven architectures and CI/CD workflows (GitLab CI). Write clean, production-grade code in Python (Scala is a bonus … Engineers and Associates, and lead technical discussions and design sessions. Key Requirements Why Holland & Barrett? Must-Have: Strong experience with AWS services: Glue, Lambda, S3, Athena, Step Functions, EventBridge, EMR, EKS, RDS, Redshift, DynamoDB. Strong Python development skills. Proficient with Docker, containerization, and virtualization. Hands-on experience with CI/CD, especially GitLab CI. Solid experience with Infrastructure as More ❯
performance-tuning skills. Preferred Qualifications Strong communication skills and demonstrated ability to engage with business stakeholders and product teams. Experience in data modeling , data warehousing (e.g., Snowflake , AWS Glue , EMR , Apache Spark ), and working with data pipelines . Leadership experience—whether technical mentorship, team leadership, or managing critical projects. Familiarity with Infrastructure as Code (IaC) tools like Terraform , CloudFormation More ❯
in SQL. Familiarity with one scripting language, preferably Python. Experience with Apache Airflow & Apache Spark. Solid understanding of cloud data services: AWS services such as S3, Athena, EC2, RedShift, EMR (ElasticMapReduce), EKS, RDS (Relational Database Services) and Lambda. Nice to have: Understanding of ML Models. Understanding of containerisation and orchestration technologies like Docker/Kubernetes. Relevant knowledge or More ❯
in SQL. Familiarity with one scripting language, preferably Python. Experience with Apache Airflow & Apache Spark. Solid understanding of cloud data services: AWS services such as S3, Athena, EC2, RedShift, EMR (ElasticMapReduce), EKS, RDS (Relational Database Services) and Lambda. Nice to have Understanding of ML Models. Understanding of containerisation and orchestration technologies like Docker/Kubernetes. Relevant knowledge or More ❯
Would you like to work on one of the world's largest transactional distributed systems? How about working with customers and peers from the entire range of Amazon's business on cool new features? Whether you're passionate about building highly scalable and reliable systems or a software developer who likes to solve business problems, Selling Partner Services … 1. Design/implement automation and manage our massive data infrastructure to scale for the analytics needs of case management. 2. Build solutions to achieve BAA(Best At Amazon) standards for system efficiency, IMR efficiency, data availability, consistency & compliance. 3. Enable efficient data exploration, experimentation of large datasets on our data platform and implement data access control mechanisms … business intelligence engineer, data scientist) with a track record of manipulating, processing, and extracting value from large datasets PREFERRED QUALIFICATIONS - Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions - Experience building large-scale, high-throughput, 24x7 data systems - Experience with big data technologies such as: Hadoop, Hive, Spark, EMRMore ❯
data pipelines and architectures on the Amazon Web Services (AWS) cloud platform. Skilled in scalable, reliable, and efficient data solutions, often using AWS services like S3, Redshift, EMR, Glue, and Kinesis. This involves designing ETL processes, ensuring data security, and collaborating with other teams for data analysis and business requirements Key Skills: AWS Services: Strong understanding and … experience with AWS services like S3 , Redshift, EMR, Glue , Kinesis, and Lambda. Programming Languages: Proficiency in programming languages like Python or Java, used for designing and building data pipelines. SQL: Knowledge of SQL for querying and manipulating data in relational databases. ETL Processes: Experience with ETL tools and techniques. Cloud Computing: Familiarity with cloud computing concepts and principles. Data More ❯
for different use cases. Develop data models and Data Lake designs around stated use cases to capture KPIs and data transformations. Identify relevant AWS services - on AmazonEMR, Redshift, Athena, Glue, Lambda, to design an architecture that can support client workloads/use-cases; evaluate pros/cons among the identified options to arrive at a recommended More ❯
for different use cases. Develop data models and Data Lake designs around stated use cases to capture KPIs and data transformations. Identify relevant AWS services – on AmazonEMR, Redshift, Athena, Glue, Lambda, to design an architecture that can support client workloads/use-cases; evaluate pros/cons among the identified options to arrive at a recommended More ❯
least 10 years of experience in data engineering, data architecture, or software engineering. Proficiency in Python and SQL Proficient in AWS data services such as S3, Glue, Athena, Redshift, EMR, Kinesis, Lambda, etc. Strong knowledge of data lake concepts, architectures, and design patterns. Experience in building and managing data pipelines using tools such as Airflow, Spark, Kinesis etc. Experience More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
PA Consulting
ll have experience working in teams to design, build, and maintain large scale data solutions and applications using AWS data and analytics services (or open-source equivalent) such as EMR, Glue, RedShift, Kinesis, Lambda, DynamoDB. Your team members will look to you as a trusted expert and will expect you to define the end-to-end software development lifecycle More ❯
Business Intelligence Engineer, Japan Operations The Amazon ARTS (APEX-RoW Technology Solutions) team is looking for a Business Intelligence Engineer to optimize in one of the world's largest and most complex data warehouse environments. You are expected being passionate about working with huge database and someone who loves to bring datasets together to create dashboards and business … and database infrastructure across RoW; 2) help business stakeholders to provide data pipeline solutions with Amazon/ARTS internal tech products and AWS services (e.g. Redshift, SageMaker, EMR, ETL tools, data lake); 3) develop scripts for maintenance automation; 4) execute the tech implementation with a clear milestone planning and lead it to smooth launch; 5) help business … India, Japan, Australia, Brazil, Mexico, Singapore, Turkey and MENA. BASIC QUALIFICATIONS - Experience with scripting language (e.g., Python, Java, or R) - Experience with AWS tools and technologies (Redshift, S3, SageMaker, EMR, Lambda) - Business level communication skills in Japanese (listening, speaking, reading, writing) - Ability to read and write in English (ex. e-mail, slack communication) - Interested in working with diverse team More ❯
passionate Data Engineer to drive innovations in translation analytics space to create the data pipelines handling large volume data and help our customer's to analyze and understand Amazon Translation coverage across the languages.We support Translation Services in making data-driven decisions by providing easy access to data and self-serve analytics. We work closely with internal stakeholders … essential Translation Services metrics. If you are not sure that every qualification on the list above describes you exactly, we'd still love to hear from you! At Amazon, we value people with unique backgrounds, experiences, and skillsets. If you’re passionate about this role and want to make an impact on a global scale, please apply! BASIC … SQL - Experience in at least one modern scripting or programming language, such as Python, Java, Scala, or NodeJS PREFERRED QUALIFICATIONS - Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions - Experience with non-relational databases/data stores (object storage, document or key-value stores, graph databases, column-family databases) - Bachelor More ❯