Architecture, or similar roles. Strong programming skills in Python/Java/Scala. Expert in SQL and performance tuning for large datasets. Hands-on experience with Big Data ecosystems Hadoop, Spark, Kafka, Hive, HBase, etc. Strong experience with Cloud platforms (AWS/Azure/GCP) and services like: AWS: S3, Glue, EMR, Redshift, Lambda, Kinesis Azure: Data Factory, Synapse More ❯
of the following: Python, SQL, Java Commercial experience in client-facing projects is a plus, especially within multi-disciplinary teams Deep knowledge of database technologies: Distributed systems (e.g., Spark, Hadoop, EMR) RDBMS (e.g., SQL Server, Oracle, PostgreSQL, MySQL) NoSQL (e.g., MongoDB, Cassandra, DynamoDB, Neo4j) Solid understanding of software engineering best practices - code reviews, testing frameworks, CI/CD, and More ❯
and data models. Required Skills & Experience Strong experience with SQL, relational databases, and NoSQL systems.Proficiency in Python, Scala, or Java for data engineering tasks.Experience with big data tools (Spark, Hadoop, Kafka, FlinkKnowledge of ETL/ELT tools (Airflow, dbt, Glue, InformaticaFamiliarity with data warehousing (Snowflake, BigQuery, RedshiftUnderstanding of data modeling, normalization, and data governance.Experience with Git, CI/CD More ❯
Bronxville, New York, United States Hybrid/Remote Options
Global Parcel Delivery
with data pipeline and workflow management tools (e.g., Apache Airflow, Luigi). Familiarity with cloud platforms such as AWS, Azure, or Google Cloud. Experience with big data technologies like Hadoop, Spark, or Kafka is a plus. Solid programming skills in Python, Java, or Scala. Understanding of data warehousing concepts and architectures. Excellent problem-solving skills and attention to detail. More ❯
ETL processes. Proficiency in Python. Experience with cloud platforms (AWS, Azure, or GCP). Knowledge of data modelling, warehousing, and optimisation. Familiarity with big data frameworks (e.g. Apache Spark, Hadoop). Understanding of data governance, security, and compliance best practices. Strong problem-solving skills and experience working in agile environments. Desirable: Experience with Docker/Kubernetes, streaming data (Kafka More ❯
data modelling, warehousing, and performance optimisation. Proven experience with cloud platforms (AWS, Azure, or GCP) and their data services. Hands-on experience with big data frameworks (e.g. Apache Spark, Hadoop). Strong knowledge of data governance, security, and compliance. Ability to lead technical projects and mentor junior engineers. Excellent problem-solving skills and experience in agile environments. Desirable: Experience More ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Hands on coding experience, such as Python Reporting tools More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (Mongo, Cassandra) 2+ years More ❯
data modelling tools, data warehousing, ETL processes, and data integration techniques. · Experience with at least one cloud data platform (e.g. AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark). · Strong knowledge of data workflow solutions like Azure Data Factory, Apache NiFi, Apache Airflow etc · Good knowledge of stream and batch processing solutions like Apache Flink, ApacheMore ❯
experience within either Flask, Tornado or Django, Docker Experience working with ETL pipelines is desirable e.g. Luigi, Airflow or Argo Experience with big data technologies, such as Apache Spark, Hadoop, Kafka, etc. Data acquisition and development of data sets and improving data quality Preparing data for predictive and prescriptive modelling Reporting tools (e.g. Tableau, PowerBI, Qlik) GDPR and Government More ❯
least one major cloud provider (AWS, Azure, or GCP Strong experience building cloud data lakes, warehouses, and streaming architectures. Proficiency with data processing tools such as Spark, Databricks, Snowflake, Hadoop, or similar. Strong knowledge of ETL/ELT frameworks, API integration, and workflow orchestration tools (Airflow, Azure Data Factory, AWS Glue, etc. Deep understanding of relational and NoSQL databases More ❯
winter garden, florida, united states Hybrid/Remote Options
Baptist Health South Florida
modern data platforms to properly design and implement data pipeline solutions. Experience using analytic SQL, working with traditional relational databases and/or distributed systems such as AWS S3, Hadoop/Hive, Redshift. Provide production support and adhering to the defined SLA(s). Strong technical skills and demonstrated ability to be detail-oriented. Good understanding of data security More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (Mongo, Cassandra) 2+ years More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (Mongo, Cassandra) 2+ years More ❯
including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (Mongo, Cassandra) 2+ years More ❯
Elizabeth, New Jersey, United States Hybrid/Remote Options
ALTA IT Services
of experience may be considered in lieu of a degree. Proficiency in programming languages like Python or Java, strong SQL skills, and knowledge of big data tools like ApacheHadoop, Spark, or Kafka. Experience with cloud platforms (AWS, Azure, GCP) and data warehousing solutions (Snowflake, Redshift, BigQuery) Self-driven and have demonstrated the ability to work independently with minimum More ❯
providers, preferably Azure (AWS/GCP also valued). Experience with modern data lakehouse concepts and architectures. Experience ingesting and transforming data from sources such as Kafka, REST APIs, Hadoop, etc. Experience with data pipeline orchestration tools (Azure Data Factor or similar). Experience implementing ETL/ELT pipelines using modern data lakehouse solutions. Experience with data quality, observability More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Michael Page
Experience within the Insurance industry Strong proficiency in programming languages such as Python, Java, or Scala. Experience with cloud platforms like Azure. Knowledge of big data technologies such as Hadoop, Spark, or Kafka. Proficiency in SQL and database management systems. Familiarity with data warehousing concepts and tools. Ability to work collaboratively with cross-functional teams. A solid understanding of More ❯
London, South East, England, United Kingdom Hybrid/Remote Options
Michael Page Technology
Experience within the Insurance industry Strong proficiency in programming languages such as Python, Java, or Scala. Experience with cloud platforms like Azure. Knowledge of big data technologies such as Hadoop, Spark, or Kafka. Proficiency in SQL and database management systems. Familiarity with data warehousing concepts and tools. Ability to work collaboratively with cross-functional teams. A solid understanding of More ❯
analytical thinking capabilities. Experience working in enterprise environments with cross-functional teams. Preferred Experience Experience with cloud data environments (AWS, Azure, or Google Cloud Familiarity with big data platforms (Hadoop, Spark, Databricks Background in industries such as finance, healthcare, government, or manufacturing. Experience with model governance, validation, and lifecycle management. Education Bachelor's or Master's degree in Data More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯
including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 5+ years experience with Distributed data/computing tools (MapReduce, Hadoop, Hive, EMR, Kafka, Spark, Gurobi, or MySQL) 4+ year experience working on real-time data and streaming applications 4+ years of experience with NoSQL implementation (Mongo, Cassandra) 4+ years More ❯