a Data Engineer for Cloud Data Lake activities, especially in high-volume data processing frameworks, ETL development using distributed computing frameworks like Apache Spark, Hadoop, Hive. Experience optimizing database performance, scalability, data security, and compliance. Experience with event-based, micro-batch, and batched high-volume, high-velocity transaction and More ❯
infrastructure-as-code (e.g., Terraform, CloudFormation), CI/CD pipelines, and monitoring (e.g., CloudWatch, Datadog). Familiarity with big data technologies like Apache Spark, Hadoop, or similar. ETL/ELT tools and creating common data sets across on-prem (IBMDatastage ETL) and cloud data stores Leadership & Strategy: Lead Data More ❯
Service Catalogue, Cloud Formation, Lake Formation, SNS, SQS, Event Bridge Language & Scripting: Python and Spark ETL: DBT Good to Have: Airflow, Snowflake, Big Data (Hadoop), and Teradata Responsibilities: Serve as the primary point of contact for all AWS related data initiatives and projects. Responsible for leading a team of More ❯
Experience in commodities markets or broader financial markets. Knowledge of quantitative modeling, risk management, or algorithmic trading. Familiarity with big data technologies like Kafka, Hadoop, Spark, or similar. Why Work With Us? Impactful Work: Directly influence the profitability of the business by building technology that drives trading decisions. Innovative More ❯
Experience in commodities markets or broader financial markets. Knowledge of quantitative modeling, risk management, or algorithmic trading. Familiarity with big data technologies like Kafka, Hadoop, Spark, or similar. Why Work With Us? Impactful Work: Directly influence the profitability of the business by building technology that drives trading decisions. Innovative More ❯
Account team within Services group (TTS) and is responsible for building a scalable, high-performance data platform on Big Data technologies (Spark, Scala, Hive, Hadoop) along with Kafka/Java and AI technologies to support core account data needs across multiple lines of businesses. As a tenant on the More ❯
Python and SQL programming languages. Hands-on experience with cloud platforms like AWS, GCP, or Azure, and familiarity with big data technologies such as Hadoop or Spark. Experience working with relational databases and NoSQL databases. Strong knowledge of data structures, data modelling, and database schema design. Experience in supporting More ❯
architectures Proficiency in writing and optimizing SQL Knowledge of AWS services including S3, Redshift, EMR, Kinesis and RDS Experience with Open Source Data Technologies (Hadoop, Hive, Hbase, Pig, Spark, etc.) Ability to write code in Python, Ruby, Scala or other platform-related Big data technology Knowledge of professional software More ❯
or related industries. Certification in relevant areas (e.g., AWS Certified Data Analytics, Google Data Analytics Professional Certificate). Familiarity with big data technologies (e.g., Hadoop, Spark) and cloud platforms (e.g., AWS, Azure). Experience with data visualization design principles and storytelling techniques. Knowledge of agile methodologies and project management. More ❯
science, mathematics, or a related quantitative field - Experience with scripting languages (e.g., Python, Java, R) and big data technologies/languages (e.g. Spark, Hive, Hadoop, PyTorch, PySpark) PREFERRED QUALIFICATIONS - Master's degree, or Advanced technical degree - Knowledge of data modeling and data pipeline design - Experience with statistical analysis, co More ❯
with unstructured datasets. Engineering best practices and standards. Experience with data warehouse software (e.g. Snowflake, Google BigQuery, Amazon Redshift). Experience with data tools: Hadoop, Spark, Kafka, etc. Code versioning (Github integration and automation). Experience with scripting languages such as Python or R. Working knowledge of message queuing More ❯
contributions to the delivery process, manage tasks, and update teams on progress. Skills & Experience: Proven experience as a Data Engineer with expertise in Databricks, Hadoop/Spark. Strong programming skills in Python, Scala, or SQL, with knowledge of CI/CD platforms. Proficiency with distributed computing frameworks and cloud More ❯
AWS Databases: MSSQL, PostgreSQL, MySQL, NoSQL Cloud: AWS (preferred), with working knowledge of cloud-based data solutions Nice to Have: Experience with graph databases, Hadoop/Spark, or enterprise data lake environments What You’ll Bring Strong foundation in computer science principles (data structures, algorithms, etc.) Experience building enterprise More ❯
london, south east england, united kingdom Hybrid / WFH Options
Randstad Digital UK
AWS Databases: MSSQL, PostgreSQL, MySQL, NoSQL Cloud: AWS (preferred), with working knowledge of cloud-based data solutions Nice to Have: Experience with graph databases, Hadoop/Spark, or enterprise data lake environments What You’ll Bring Strong foundation in computer science principles (data structures, algorithms, etc.) Experience building enterprise More ❯
e.g. Tensorflow, MXNet, scikit-learn). Knowledge of software engineering practices (coding practices to DS, unit testing, version control, code review). Experience with Hadoop (especially the Cloudera and Hortonworks distributions), other NoSQL (especially Neo4j and Elastic), and streaming technologies (especially Spark Streaming). Deep understanding of data manipulation More ❯
Java, C++) and experience with DevOps practices (CI/CD). Familiarity with containerization (Docker, Kubernetes), RESTful APIs, microservices architecture, and big data technologies (Hadoop, Spark, Flink). Knowledge of NoSQL databases (MongoDB, Cassandra, DynamoDB), message queueing systems (Kafka, RabbitMQ), and version control systems (Git). Preferred Skills: Experience More ❯
would be great if you also had: Experience of configuring and using ETL platforms such as SSIS, AWS or Azure Data Factory Experience of Hadoop and Jenkins Azure Certified AWS Certified Familiarity with Java Knowledge of DevOps practices, including CI/CD pipelines What we do for you: At More ❯
with distributed systems as it pertains to data storage and computing Experience with Redshift, Oracle, NoSQL etc. Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Bachelor's degree PREFERRED QUALIFICATIONS Experience working on and delivering end to end projects independently Experience providing technical leadership and mentoring More ❯
Apache Airflow, DBT, or Kafka • Knowledge of cloud data services (AWS S3/Glue/Redshift, GCP BigQuery, Azure Data Factory) • Exposure to Spark, Hadoop, or other big data frameworks • Personal or academic data engineering projects ⸻ 🎁 Perks & Benefits : • 1:1 mentorship with senior data engineers • Live experience with production More ❯
Apache Airflow, DBT, or Kafka • Knowledge of cloud data services (AWS S3/Glue/Redshift, GCP BigQuery, Azure Data Factory) • Exposure to Spark, Hadoop, or other big data frameworks • Personal or academic data engineering projects ⸻ 🎁 Perks & Benefits : • 1:1 mentorship with senior data engineers • Live experience with production More ❯
Apache Airflow, DBT, or Kafka • Knowledge of cloud data services (AWS S3/Glue/Redshift, GCP BigQuery, Azure Data Factory) • Exposure to Spark, Hadoop, or other big data frameworks • Personal or academic data engineering projects ⸻ 🎁 Perks & Benefits : • 1:1 mentorship with senior data engineers • Live experience with production More ❯
platform implementation Hands on experience with building data or machine learning pipeline Experience with one or more relevant tools (SQL, PostgreSQL, Data lake, DynamoDB, Hadoop) Current experience with hands-on implementation and performance tuning of Kinesis, Kafka, Spark or similar implementations Masters or PhD in Computer Science, Physics, Engineering More ❯
Richmond, North Yorkshire, Yorkshire, United Kingdom
Datix Limited
programming languages, specifically Python and SQL. Expertise in data management, data architecture, and data visualization techniques. Experience with data processing frameworks like Apache Spark, Hadoop, or Flink. Strong understanding of database systems (SQL and NoSQL) and data warehousing technologies. Familiarity with cloud computing platforms (AWS, Azure) and data security More ❯
will be deployed You have experience in database technologies including writing complex queries against their (relational and non-relational) data stores (e.g. Postgres, ApacheHadoop, Elasticsearch, Graph databases), and designing the database schemas to support those queries You have a good understanding of coding best practices & design patterns and More ❯
such as Pandas, NumPy, and SQLAlchemy. Extensive experience with Dash framework for building web applications. In-depth knowledge of Impala or other SQL-on-Hadoop query engines. Understanding of web development concepts (HTML, CSS, JavaScript). Proficiency in data visualization libraries (Plotly, Seaborn). Solid understanding of database design More ❯