with NoSQL data persistence - DynamoDB, MongoDB, etc. DevOps mindset - knowing how to automate development and operational tasks Big data or data science background: ML, Apache Spark, ApacheHive, machine learning #J-18808-Ljbffr More ❯
Working experience in Palantir Foundry platform is must • Experience designing and implementing data analytics solutions on enterprise data platforms and distributed computing (Spark/Hive/Hadoop preferred). • Proven track record of understanding and transforming customer requirements into a best-fit design and architecture. • Demonstrated experience in end More ❯
London, England, United Kingdom Hybrid / WFH Options
Lemongrass Consulting
and migration of these data warehouses to modern cloud data platforms. Deep understanding and hands-on experience with big data technologies like Hadoop, HDFS, Hive, Spark and cloud data platform services. Proven track record of designing and implementing large-scale data architectures in complex environments. CICD/DevOps experience More ❯
Lambda usw. Umfangreiches Verständnis von Datenbankkonzepten und -technologien, einschließlich relationaler Datenbanken und NoSQL-Datenbanken. Praktische Erfahrung mit ETL-/ELT-Tools und -Frameworks wie Apache Spark, Apache Airflow, AWS Glue und/oder AWS Data Pipeline. Kenntnisse in der Datenmodellierung und -optimierung für analytische und Machine-Learning-Anwendungen. … Erfahrung mit der Programmierung in Sprachen wie Python, Scala und/oder ähnlichen Sprachen. Vertrautheit mit gängigen Big-Data-Technologien und -Konzepten wie Hadoop, Hive, MapReduce, Apache Kafka, sowie Streaming-Technologien wie Kinesis, Apache Flink und/oder ähnlichem. Kenntnisse in der Implementierung und Verwaltung von Datenpipelines More ❯
preferably in Financial Services) in building enterprise Cloud applications based on high-volume data processing frameworks, ETL development using distributed computing frameworks such as Apache Spark, Hadoop, Hive. Experience with optimizing database performance, scalability, and ensuring data security and compliance. Experience with event-based, micro-batch, and batched high More ❯
MDX, HiveQL, SparkSQL, Scala) - Experience with one or more scripting language (e.g., Python, KornShell) PREFERRED QUALIFICATIONS - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results More ❯
computer science, mathematics, or a related quantitative field - Experience with scripting languages (e.g., Python, Java, R) and big data technologies/languages (e.g. Spark, Hive, Hadoop, PyTorch, PySpark) PREFERRED QUALIFICATIONS - Master's degree, or Advanced technical degree - Knowledge of data modeling and data pipeline design - Experience with statistical analysis More ❯
analytics, or data science, with the ability to work effectively with various data types and sources. Experience using big data technologies (e.g. Hadoop, Spark, Hive) and database management systems (e.g. SQL and NoSQL). Graph Database Expertise: Deep understanding of graph database concepts, data modeling, and query languages (e.g. More ❯
lake platform and different components in the data lake such as Hadoop, Amazon S3 etc. Work on SQL technologies on Hadoop such as Spark, Hive, Impala etc Help continually improve ongoing analysis processes, optimizing or simplifying self-service support for customers Must possess strong verbal and written communication skills More ❯
pipelines Proficiency in SQL Experience with scripting languages like Python or KornShell Unix experience Troubleshooting data and infrastructure issues Preferred Qualifications Experience with Hadoop, Hive, Spark, EMR Experience with ETL tools like Informatica, ODI, SSIS, BODI, DataStage Knowledge of distributed storage and computing systems Experience with reporting and analytics More ❯
platforms. Experience in writing efficient SQL queries and implementing complex ETL transformations on big data platforms. Experience with Big Data technologies like Spark, Impala, Hive, Redshift, Kafka, etc. Experience in data quality testing; capable of writing test cases and scripts, and resolving data issues. Experience with Databricks, Snowflake, Iceberg More ❯
London, England, United Kingdom Hybrid / WFH Options
Solirius Reply
TensorFlow, XGBoost, PyTorch). Strong foundation in statistics, probability, and hypothesis testing. Experience with cloud platforms (AWS, GCP, Azure) and big data tools (Spark, Hive, Databricks, etc.) is a plus. Excellent communication and storytelling skills with the ability to explain complex concepts to non-technical stakeholders. Proven track record More ❯
London, England, United Kingdom Hybrid / WFH Options
Citi
platforms implementation Hands-on development and architecture experience in Java or Scala, Microservices Hands-on development and architecture experience in big-data ecosystem – Hadoop, Hive, Impala, NOSQL Hands-on development, architecture and leadership experience in distributed computing technologies – Spark, Flink Hands-on development, and architecture experience in caching, In More ❯
Cloud Data Lake activities. The candidate should have industry experience (preferably in Financial Services) in navigating enterprise Cloud applications using distributed computing frameworks as Apache Spark, Hadoop, Hive. Working knowledgeoptimizing database performance, scalability, ensuring data security and compliance. Education & Preferred Qualifications Bachelor’s/Master's Degree in a More ❯
MDX, HiveQL, SparkSQL, Scala) - Experience with one or more scripting language (e.g., Python, KornShell) PREFERRED QUALIFICATIONS - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results More ❯
the big 3 cloud ML stacks (AWS, Azure, GCP). Hands-on experience with open-source ETL, and data pipeline orchestration tools such as Apache Airflow and Nifi. Experience with large scale/Big Data technologies, such as Hadoop, Spark, Hive, Impala, PrestoDb, Kafka. Experience with workflow orchestration … tools like Apache Airflow. Experience with containerisation using Docker and deployment on Kubernetes. Experience with NoSQL and graph databases. Unix server administration and shell scripting experience. Experience in building scalable data pipelines for highly unstructured data. Experience in building DWH and data lakes architectures. Experience in working in cross More ❯
expertise in Cloudera Data Platform (CDP), Cloudera Manager, and Cloudera Navigator . Strong knowledge of Hadoop ecosystem and related technologies such as HDFS, YARN, Hive, Impala, Spark, and Kafka . Strong AWS services/Architecture experience with hands-on expertise in cloud-based deployments (AWS, Azure, or GCP) . More ❯
or KornShell. Knowledge of writing and optimizing SQL queries for large-scale, complex datasets. PREFERRED QUALIFICATIONS Experience with big data technologies such as Hadoop, Hive, Spark, EMR. Experience with ETL tools like Informatica, ODI, SSIS, BODI, or DataStage. We promote an inclusive culture that empowers Amazon employees to deliver More ❯
London, England, United Kingdom Hybrid / WFH Options
JR United Kingdom
field. • 3+ experience as a Software Engineer with a strong focus on Data work. • Strong proficiency in leading big data technologies (e.g., Hadoop, Spark, Hive). • Familiarity with data warehousing solutions (e.g., Redshift, BigQuery, Snowflake). • Strong problem-solving skills and attention to detail. • Excellent communication and collaboration skills. More ❯
FireHose, Lambda, and IAM roles and permissions - Experience building large-scale, high-throughput, 24x7 data systems - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience providing technical leadership and mentoring other engineers for best practices on data engineering Our inclusive culture empowers Amazonians to deliver the More ❯
London, England, United Kingdom Hybrid / WFH Options
Enigma
field. • 3+ experience as a Software Engineer with a strong focus on Data work. • Strong proficiency in leading big data technologies (e.g., Hadoop, Spark, Hive). • Familiarity with data warehousing solutions (e.g., Redshift, BigQuery, Snowflake). • Strong problem-solving skills and attention to detail. • Excellent communication and collaboration skills. More ❯
SparkSQL, Scala). Experience with one or more scripting language (e.g., Python, KornShell). PREFERRED QUALIFICATIONS Experience with big data technologies such as: Hadoop, Hive, Spark, EMR. Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results More ❯
London, England, United Kingdom Hybrid / WFH Options
Qh4 Consulting
C#) and willingness to interact with them where necessary. Exposure to Cloudera Data Platform or similar big data environments. Experience with tools such as ApacheHive, NiFi, Airflow, Azure Blob Storage, and RabbitMQ. Background in investment management or broader financial services, or a strong willingness to learn the More ❯
rapid prototyping and disciplined software development processes. Experience with Python, ML libraries (e.g. spaCy, NumPy, SciPy, Transformers, etc.), data tools and technologies (Spark, Hadoop, Hive, Redshift, SQL), and toolkits for ML and deep learning (SparkML, Tensorflow, Keras). Demonstrated ability to work on multi-disciplinary teams with diverse skillsets. More ❯
have 4+ years of relevant work experience in Analytics, Business Intelligence, or Technical Operations Master in SQL, Python, and ETL using big data tools (HIVE/Presto, Redshift) Previous experience with web frameworks for Python such as Django/Flask is a plus Experience writing data pipelines using Airflow More ❯