and cost-effectiveness. Implement data quality checks and validation rules within data pipelines. Data Transformation & Processing: Implement complex data transformations using Spark (PySpark or Scala) and other relevant technologies. Develop and maintain data processing logic for cleaning, enriching, and aggregating data. Ensure data consistency and accuracy throughout the data lifecycle. More ❯
and principles. Solid understanding of data warehousing data modeling and data integration principles. Proficiency in at least one scripting/programming language (e.g. Python Scala Java). Experience with SQL and NoSQL databases. Familiarity with data quality and data governance best practices. Strong analytical and problem-solving skills. Excellent communication More ❯
leadership role. Hands-on experience with data warehouse solutions such as Redshift, Snowflake, or BigQuery. Strong command of SQL and programming languages like Python, Scala, or Java. Familiarity with ETL/ELT tools (e.g., Airflow, Fivetran, dbt) and cloud data stacks (AWS/GCP/Azure). A deep understanding More ❯
data modeling , distributed systems , streaming architectures , and ETL/ELT pipelines . Proficiency in SQL and at least one programming language such as Python , Scala , or Java . Demonstrated experience owning and delivering complex systems from architecture through implementation. Excellent communication skills with the ability to explain technical concepts to More ❯
building and maintaining cloud-based data systems (e.g., Redshift, BigQuery, Snowflake) Strong coding skills in languages commonly used for data work (e.g., Python, Java, Scala) Deep understanding of ETL/ELT tools and workflow orchestration platforms (e.g., Airflow, Fivetran, dbt) Proficiency with SQL and solid grounding in data modeling concepts More ❯
processes . Strong SQL skills , with the ability to write optimized and scalable queries. Or, Proficiency in at least one programming language ( Python , Java, Scala, or .NET). CI/CD : Experience using CI/CD pipelines for development and deployment of data pipelines. Proficiency in Git-based workflows and More ❯
of performing architectural assessments, examining architectural alternatives, and choosing the best solution in collaboration with both IT and business stakeholders Fluent in Python, Java, Scala, or similar Object-Oriented Programming Languages Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with More ❯
Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced More ❯
advantageous Experience of knowledge of containers such as Docker and Kubernetes is advantageous Familiarity with at least one programming language (e.g. Python, Java, or Scala) Proven experience of data warehousing concepts and ETL processes Strong analytical skills and attention to detail Excellent verbal and written communication skills in English is More ❯
technical concepts for different audiences. Experience working within a Unix-based environment. Background in building distributed data systems or pipelines at scale. Knowledge of Scala, Spark, or experience supporting machine learning pipelines is a bonus. If this role interests you and you would like to find out more (or find More ❯
Domain-driven design, SOLID, DRY, SoC, etc. • Software development background of large-scale distributed systems with exposure to multiple languages (like Java, C#, Go, Scala, Python etc.) • Hands-on experience with frameworks like Spring Boot, Node.js • Experience with/exposure to frontend technology frameworks like React, Angular etc • Experience in More ❯
looking for? Experience in the design and deployment of production data pipelines from ingestion to consumption within a big data architecture, using Java, Python, Scala, Spark, SQL. Experience performing tasks such as writing scripts, extracting data using APIs, writing SQL queries etc. Ability to closely with other engineering teams to More ❯
City of London, Greater London, UK Hybrid / WFH Options
CipherTek Recruitment
Expertise in Azure DevOps and GitHub Actions. Familiarity with Databricks CLI and Databricks Job Bundle. Strong programming skills in Python and SQL; familiarity with Scala is a plus. Solid understanding of AI/ML algorithms, model training, evaluation (including hyperparameter tuning), deployment, monitoring, and governance. Experience in handling large datasets More ❯
platforms, and a very smart team You might be a good fit if you: Have 1–3 years of experience writing software (Python, Java, Scala — whatever works) Know your way around SQL and databases Have touched cloud services (AWS, GCP, etc.) Enjoy solving problems, learning fast, and working with good More ❯
you will: Design and deploy production data pipelines from ingestion to consumption within a big data architecture. Work with technologies such as Python, Java, Scala, Spark, and SQL to extract, clean, transform, and integrate data. Build scalable solutions using AWS services like EMR, Glue, Redshift, Kinesis, Lambda, and DynamoDB. Process More ❯
objectives and compliance standards Support test and deployment of new products and features Participate in code reviews 🌱 About You Expert knowledge of Java/Scala/Kotlin. Familiarity with Kotlin or willingness to learn. Industrial experience with AWS/GCP/Azure. Knowledge of common data products such as Hadoop More ❯
experience in the following skills: Relevant work experience in data science, machine learning, and business analytics Practical experience in coding languages eg. Python, R, Scala, etc.; (Python preferred) Proficiency in database technologies eg. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. pySpark, Hive, etc. Experienced working with structured More ❯
with your team. Key Skills Python in the software engineering level, including unit and integration test experience. Distributed computing knowledge covered by PySpark or Scala, can debug things in SparkUI and knows how to optimise for this purpose. AWS experience Good understanding of data modelling, change data capture and/ More ❯
help with their data presentation requirements and help deliver data for visualisation solutions, such as Power BI Key skills required: Expertise in Python, Java, Scala, or other programming languages used to build data pipelines, implement data transformations, and automate data workflows. Strong SQL experience. Strong knowledge of data visualisation tools More ❯
manage and process large datasets. Actively contribute to the continuous improvement of data operations and infrastructure. Skills and Experience Required: Proficiency in SQL, Python, Scala, and R. Experience with Big Data technologies such as Microsoft Fabric, Microsoft Synapse, Spark, and Kafka. Familiarity with database management systems including SQL Server and More ❯
TDD You have a thorough understanding of Computer Science fundamentals such as OOP, Design Patterns, Data Structures, Algorithms Other tech in the stack includes Scala, React, Spring, Oracle, Redis, Kubernetes, Docker and Linux so previous exposure to any of these would be beneficial You're collaborative with good communication skills More ❯
Requirements of the Database Engineer: Proven experience with Databricks, Azure Data Lake, and Delta Live Tables Strong programming in Python and Spark (PySpark or Scala) Solid knowledge of data modelling, warehousing, and integration concepts Comfortable working in Agile teams, with CI/CD and Azure DevOps experience Package: Salary depending More ❯
to encourage their development and strengthen the team's overall skillset. Skills Profile Strong command of JVM-based languages, with openness to Kotlin, Java, Scala, or Groovy. Experience delivering scalable, production-quality software solutions end-to-end. Focused on outcomes, with practical experience in service-oriented and microservices architectures. Hands More ❯