migration of legacy ETL workflows and data systems to cloud-native architectures, delivering measurable cost, scalability, and performance improvements. Big Data Technologies Hands-on experience with distributed computing frameworks (Hadoop, Spark, Hive, Presto). Proficiency with data lake and lakehouse architectures (Delta Lake, Apache Iceberg, Apache Hudi). Understanding of partitioning, data compaction, schema evolution, and ACID compliance. Strong More ❯
an incremental value to customers Bonus points for: Experience with cloud-based data services and infrastructure (e.g., AWS, GCP, Azure) Experience with big data technologies like Apache Spark or Hadoop Some Knowledge of data visualization and business intelligence tools like Tableau, Power BI, or Looker Experience in a data-centric role supporting machine learning initiatives, including building data pipelines More ❯
of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
scripting, PERL and JAVA, XML/JSON files, SAS , Python, AWS cloud-native technologies, S3, Athena, Redshift Experience in snowflake is an added bonus. Familiarity with the following technologies: Hadoop, Kafka, Airflow, Hive, Presto, Athena, S3, Aurora, EMR, Spark Ability to drive, contribute to, and communicate solutions to technical product challenges Ability to roll-up your sleeves and work More ❯
a Data Engineer or in a similar role. Strong proficiency in SQL and experience with relational databases (e.g., MySQL, PostgreSQL, Oracle). Familiarity with big data technologies such as Hadoop, Spark, and Kafka. Experience with data integration tools and platforms (e.g., Talend, Apache NiFi). Proficiency in programming languages such as Python, Java, or Scala. Knowledge of cloud-based More ❯
a Data Engineer or in a similar role. Strong proficiency in SQL and experience with relational databases (e.g., MySQL, PostgreSQL, Oracle). Familiarity with big data technologies such as Hadoop, Spark, and Kafka. Experience with data integration tools and platforms (e.g., Talend, Apache NiFi). Proficiency in programming languages such as Python, Java, or Scala. Knowledge of cloud-based More ❯
e.g., PostgreSQL, SQL Server, DynamoDB). • Experience with cloud data platforms such as AWS (S3, Redshift, Glue), Azure, Snowflake, or Databricks. • Familiarity with big data tools (e.g., Spark, Kafka, Hadoop). • Version control experience (e.g., Git). • Strong communication skills and the ability to translate technical concepts to non-technical stakeholders. • Comfortable working autonomously and navigating ambiguity in a More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
operational decision support and analyzing unstructured data (e.g., text, imagery). Ability to architect and maintain scalable data lakes, data warehouses, or distributed storage systems (e.g., Delta Lake, Snowflake, Hadoop, or NoSQL solutions). Demonstrated understanding of data security, privacy, and sovereignty issues, particularly in military or international environments, ensuring compliance with NATO operational and ethical standards. Experience building More ❯
on large-scale data systems. Strong programming skills in Python, SQL, and familiarity with Java/Scala a plus. Hands-on experience with big data frameworks (e.g., Spark, Flink, Hadoop) and workflow orchestration (Airflow, Prefect, Dagster). Proven experience with cloud-based data platforms (AWS, GCP, Azure) and data lake/warehouse technologies (Snowflake, BigQuery, Redshift, Delta Lake). More ❯
similar role, with a strong understanding of data architecture and pipeline design. Proficiency in programming languages such as Python, Java, or Scala. Experience with big data technologies such as Hadoop, Spark, Kafka, or similar. Strong knowledge of SQL and experience with relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra). Familiarity with cloud platforms (AWS, GCP More ❯
data modelling tools, data warehousing, ETL processes, and data integration techniques. Experience with at least one cloud data platform (e.g. AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark). Strong knowledge of data workflow solutions like Azure Data Factory, Apache NiFi, Apache Airflow etc Good knowledge of stream and batch processing solutions like Apache Flink, ApacheMore ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Solirius Reply
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
communication skills We're excited if you have 7+ years of experience delivering multi tier, highly scalable, distributed web applications Experience working with Distributed computing frameworks knowledge: Hive/Hadoop, Apache Spark, Kafka, Airflow Working with programming languages Python , Java, SQL. Working on building ETL (Extraction Transformation and Loading) solution using PySpark Experience in SQL/NoSQL database design More ❯
engineering Experience coding in SQL/Python, with solid CS fundamentals including data structure and algorithm design Hands-on implementation experience working with a combination of the following technologies: Hadoop, Map Reduce, Kafka, Hive, Spark, SQL and NoSQL data warehouses Knowledge of Data Warehousing, design, implementation and optimization Practice in Data Quality testing, automation and results visualization Competency in More ❯
offerings Experience working with teams located in multiple locations across the globe Knowledge and understanding of various Data repositories, Databases, ETL and BI tools and data assets hosted on Hadoop or cloud platforms Agile development experience & collaborating with data engineering teams Familiarity with Spark, Scala and distributed data processing platforms will be a big advantage Ability to work with More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯