AWS Data Engineer - (Python/PySpark/Aws Services/Unit testing/CI/CD/Gitlab/Banking)

I am hiring for AWS Data Engineer

Location: Glasgow 2–3 days per weekly Onsite

Job Description

We are looking for an experienced AWS Data Engineer with strong hands-on coding skills and expertise in designing scalable cloud-based data solutions. The ideal candidate will be proficient in Python, PySpark, and core AWS services, with a strong background in building robust data pipelines and cloud-native architectures.

Key Responsibilities

  • Design, develop, and maintain scalable data pipelines and ETL workflows using AWS services.
  • Implement data processing solutions using PySpark and AWS Glue.
  • Build and manage infrastructure as code using CloudFormation.
  • Develop serverless applications using Lambda, Step Functions, and S3.
  • Perform data querying and analysis using Athena.
  • Support Data Scientists in model operationalization using SageMaker.
  • Ensure secure data handling using IAM, KMS, and VPC configurations.
  • Containerize applications using ECS.
  • Write clean, testable Python code with strong unit testing practices.
  • Use GitLab for version control and CI/CD.

Key Skills

Python, PySpark, S3, Lambda, Glue, Step Functions, Athena, SageMaker, VPC, ECS, IAM, KMS, CloudFormation, GitLab

Job Details

Company
GIOS Technology
Location
Glasgow, UK
Employment Type
Full-time
Posted