Our team values continuous learning, knowledge sharing, and creating inclusive solutions that make a difference. Key Responsibilities Support customers with big data services including Apache Spark, Hive, Presto, and other Hadoop ecosystem components Develop and share technical solutions through various communication channels Contribute to improving support processes and … work week schedule, which may include weekends on rotation. BASIC QUALIFICATIONS - Good depth of understanding in Hadoop Administration, support and troubleshooting (Any two applications: Apache Spark, ApacheHive, Presto, Map-Reduce, Zookeeper, HBASE, HDFS and Pig.) - Good understanding of Linux and Networking concepts - Intermediate programming/scripting More ❯
computer science, mathematics, finance or equivalent quantitative field - Experience with scripting languages (e.g., Python, Java, R) and big data technologies/languages (e.g. Spark, Hive, Hadoop, PyTorch, PySpark) to build and maintain data pipelines and ETL processes - Demonstrate proficiency in SQL, data analysis, and data visualization tools like Amazon More ❯
Ruby PREFERRED QUALIFICATIONS - 3+ years of analyzing and interpreting data with Redshift, Oracle, NoSQL etc. experience - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions Our inclusive culture empowers More ❯
Proficiency in writing and optimizing SQL Knowledge of AWS services including S3, Redshift, EMR, Kinesis and RDS Experience with Open Source Data Technologies (Hadoop, Hive, Hbase, Pig, Spark, etc.) Ability to write code in Python, Ruby, Scala or other platform-related Big data technology Knowledge of professional software engineering More ❯
Proficiency in writing and optimizing SQL Knowledge of AWS services including S3, Redshift, EMR, Kinesis and RDS. Experience with Open Source Data Technologies (Hadoop, Hive, Hbase, Pig, Spark, etc.) Ability to write code in Python, Ruby, Scala or other platform-related Big data technology Knowledge of professional software engineering More ❯
to streamline data workflows and reduce manual interventions. Must have: AWS, ETL, EMR, GLUE, Spark/Scala, Java, Python. Good to have: Cloudera - Spark, Hive, Impala, HDFS, Informatica PowerCenter, Informatica DQ/DG, Snowflake Erwin. Qualifications: Bachelor's or Master's degree in Computer Science, Data Engineering, or a More ❯
SQL, DDL, MDX, HiveQL, SparkSQL, Scala) Experience with one or more scripting languages (e.g., Python, KornShell) Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Experience with any ETL tool like Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results More ❯
distributed systems as it pertains to data storage and computing Experience with Redshift, Oracle, NoSQL etc. Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Bachelor's degree PREFERRED QUALIFICATIONS Experience working on and delivering end to end projects independently Experience providing technical leadership and mentoring other More ❯
Experience in working with data visualization tools Experience in GCP tools – Cloud Function, Dataflow, Dataproc and Bigquery Experience in data processing framework – Beam, Spark, Hive, Flink GCP data engineering certification is a merit Have hands on experience in Analytical tools such as powerBI or similar visualization tools Exhibit understanding More ❯
Experience as a Data Engineer for Cloud Data Lake activities, especially in high-volume data processing frameworks, ETL development using distributed computing frameworks like Apache Spark, Hadoop, Hive. Experience optimizing database performance, scalability, data security, and compliance. Experience with event-based, micro-batch, and batched high-volume, high-velocity More ❯
Experience in working with data visualization tools Experience in GCP tools – Cloud Function, Dataflow, Dataproc and Bigquery Experience in data processing framework – Beam, Spark, Hive, Flink GCP data engineering certification is a merit Have hands on experience in Analytical tools such as powerBI or similar visualization tools Exhibit understanding More ❯
the big 3 cloud ML stacks (AWS, Azure, GCP). Hands-on experience with open-source ETL, and data pipeline orchestration tools such as Apache Airflow and Nifi. Experience with large scale/Big Data technologies, such as Hadoop, Spark, Hive, Impala, PrestoDb, Kafka. Experience with workflow orchestration … tools like Apache Airflow. Experience with containerisation using Docker and deployment on Kubernetes. Experience with NoSQL and graph databases. Unix server administration and shell scripting experience. Experience in building scalable data pipelines for highly unstructured data. Experience in building DWH and data lakes architectures. Experience in working in cross More ❯
SparkSQL, Scala). Experience with one or more scripting language (e.g., Python, KornShell). PREFERRED QUALIFICATIONS Experience with big data technologies such as: Hadoop, Hive, Spark, EMR. Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results More ❯
independently while also thriving in a collaborative team environment. Experience with GenAI/LLMs projects. Familiarity with distributed data/computing tools (e.g., Hadoop, Hive, Spark, MySQL). Background in financial services, including banking or risk management. Knowledge of capital markets and financial instruments, along with modelling expertise. If More ❯
MDX, HiveQL, SparkSQL, Scala) Experience with one or more scripting language (e.g., Python, KornShell) PREFERRED QUALIFICATIONS Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Experience as a data engineer or related specialty (e.g., software engineer, business intelligence engineer, data scientist) with a track record of manipulating More ❯
Lincoln, Lincolnshire, United Kingdom Hybrid / WFH Options
Adecco
CRITERIA Degree in Computer Science, Information Systems, or a related field. or a combination of education and relevant experience Query languages e.g. SQL, Java, Hive, R Data Management technologies e.g. ETL tools, data integration platforms Proven experience as a Data Architect, Data Engineer (or a related role, with a More ❯
experience working with relational and non-relational databases (e.g. Snowflake, BigQuery, PostgreSQL, MySQL, MongoDB). Hands-on experience with big data technologies such as Apache Spark, Kafka, Hive, or Hadoop. Proficient in at least one programming language (e.g. Python, Scala, Java, R). Experience deploying and maintaining cloud More ❯
East London, London, United Kingdom Hybrid / WFH Options
McGregor Boyall Associates Limited
and training techniques . Experience deploying models in production environments. Nice to Have: Experience in GenAI/LLMs Familiarity with distributed computing tools (Hadoop, Hive, Spark). Background in banking, risk management, or capital markets . Why Join? This is a unique opportunity to work at the forefront of More ❯
and building ETL pipelines - Experience with SQL - Experience mentoring team members on best practices PREFERRED QUALIFICATIONS - Experience with big data technologies such as Hadoop, Hive, Spark, EMR - Experience operating large data warehouses Amazon is an equal opportunities employer. We believe passionately that employing a diverse workforce is central to More ❯
computing Big Data & Distributed Systems Knowledge: Strong understanding of how distributed systems function, particularly in large-scale data environments Hands-on experience with Hadoop, ApacheHive, or similar big data technologies Apache Spark Expertise (Mandatory): Experience in batch data processing using Apache Spark Writing and optimizing More ❯
computing Big Data & Distributed Systems Knowledge: Strong understanding of how distributed systems function, particularly in large-scale data environments Hands-on experience with Hadoop, ApacheHive, or similar big data technologies Apache Spark Expertise (Mandatory): Experience in batch data processing using Apache Spark Writing and optimizing More ❯
computing Big Data & Distributed Systems Knowledge: Strong understanding of how distributed systems function, particularly in large-scale data environments Hands-on experience with Hadoop, ApacheHive, or similar big data technologies Apache Spark Expertise (Mandatory): Experience in batch data processing using Apache Spark Writing and optimizing More ❯
computing Big Data & Distributed Systems Knowledge: Strong understanding of how distributed systems function, particularly in large-scale data environments Hands-on experience with Hadoop, ApacheHive, or similar big data technologies Apache Spark Expertise (Mandatory): Experience in batch data processing using Apache Spark Writing and optimizing More ❯
computing Big Data & Distributed Systems Knowledge: Strong understanding of how distributed systems function, particularly in large-scale data environments Hands-on experience with Hadoop, ApacheHive, or similar big data technologies Apache Spark Expertise (Mandatory): Experience in batch data processing using Apache Spark Writing and optimizing More ❯