Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citi
are Accurate, Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other … practices, and escalating, managing and reporting control issues with transparency. Skills & Qualifications: Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage, capacity planning, investigating. More ❯
computation needssuch as Hbase, CloudBase/Acumulo, Big Table. Shall have demonstrated work experience with the Map Reduce programming model and technologiessuch as Hadoop, Hive, Pig. Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS). Shall have demonstrated work experience with Serialization such as JSON More ❯
presentation, and interpersonal skills. Experience working across cross-functional teams. Experience managing competing priorities and simultaneous projects. Preferred Qualifications Experience in SQL (Oracle, Vertica, Hive, MySQL, etc.). Experience in Identity and Access Management operations. Practical experience of understanding and interest in current and emerging technologies demonstrated through training More ❯
London, England, United Kingdom Hybrid / WFH Options
AlphaSights
automate data ingestion, transformation, and orchestration across systems. Own data operations infrastructure: Manage and optimise key data infrastructure components within AWS, including Amazon Redshift, Apache Airflow for workflow orchestration, and other analytical tools. You will be responsible for ensuring the performance, reliability, and scalability of these systems to meet … leveraging AWS data services. Strong proficiency in DataOps methodologies and tools, including experience with CI/CD pipelines, containerized applications, and workflow orchestration using Apache Airflow. Familiar with ETL frameworks, and bonus experience with Big Data processing (Spark, Hive, Trino), and data streaming. Proven track record – You’ve More ❯
factor app development standards Experience building modern enterprise applications and deploying to public or private clouds including AWS Experience in distributed cache systems like Apache Ignite or Redis Experience in big data platforms and technologies such as Hadoop, Hive, HDFS, Presto/Starburst, Spark, and Kafka Experience in More ❯
insight initiatives. BASIC QUALIFICATIONS - Bachelor's degree in computer science or equivalent - 3+ years of big data technologies such as AWS, Hadoop, Spark, Pig, Hive, Lucene/SOLR or Storm/Samza experience - Experience with diverse data formats: Parquet, JSON, big data formats, and table formats like ApacheMore ❯
database architecture and performance, implementing DevSecOps practices, and building CI/CD pipelines using Python, Bash, and Terraform. Preferred candidates will have experience with Apache Spark, Apache Nifi, data governance, and ETL standardization. Familiarity with Glue, Hive, and Iceberg or similar technologies is a plus. Tasks Performed … queries. • Plan and execute large-scale data migrations. • Improve database performance through architecture and tuning. • Create and maintain data flows using ETL tools like Apache Nifi. • Manage infrastructure as code using Python, Bash, and Terraform. • Integrate security into development and deployment workflows. • Build and support automated CI/CD … experience in large-scale data migration efforts. • Demonstrated experience with database architecture, performance design methodologies, and system-tuning recommendations. Preference for familiarity with Glue, Hive, and Iceberg or similar. • Demonstrated experience with Python, Bash, and Terraform. • Demonstrated experience with DevSecOps solutions and tools. • Demonstrated experience implementing CI/CD More ❯
Midvale, UT) Program, debug, and test applications and frameworks to be used and shared on the Big Data stack, including Hadoop MapReduce, Spark, Pig, Hive, and various streaming technologies in a development environment. Responsible for frameworks More ❯
forensics, log analysis) Experience interpreting information from multiple sources and working with data sets Knowledge with database tools/systems such as Hbase, SQL, Hive Query Language Preferred Qualifications Coding proficiency in Python, PHP, and/or C++, or similar high level languages About Meta Meta builds technologies that More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citigroup Inc
are Accurate, Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other … Application Support role. Hands-on experience in supporting applications built in Hadoop. Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Experienced in Linux Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage More ❯
Associate Director, Data Science and Innovation Are you a data science leader passionate about leveraging AI and machine learning to drive innovation in financial markets and transaction banking? Standard Chartered Bank is on an ambitious journey to embed cutting-edge More ❯
TensorFlow, PyTorch, or JAX. Knowledge of data analytics concepts, including data warehouse technical architectures, ETL and reporting/analytic tools and environments (such as Apache Beam, Hadoop, Spark, Pig, Hive, MapReduce, Flume). Customer facing experience of discovery, assessment, execution, and operations. Demonstrated excellent communication, presentation, and problem More ❯
TensorFlow, PyTorch, or JAX. Knowledge of data analytics concepts, including data warehouse technical architectures, ETL and reporting/analytic tools and environments (such as Apache Beam, Hadoop, Spark, Pig, Hive, MapReduce, Flume). Customer facing experience of discovery, assessment, execution, and operations. Demonstrated excellent communication, presentation, and problem More ❯
YAML Python (advanced, Pydantic experience a bonus) SQL PySpark Delta Lake Bash (CLI usage and scripting) Git Markdown Scala (bonus) Azure SQL Server as HIVE Metastore (bonus) Technologies Azure Databricks Apache Spark Delta Tables Data processing with Python PowerBI (Data Ingestion/Integration) JIRA If this role suits More ❯
proactively resolve data inconsistencies. · Participate in troubleshooting and performance tuning of ETL jobs and workflows. Required Skills & Qualifications: · Proven experience with Talend, Python, and Apache Spark. · Strong understanding of relational databases and Big Data ecosystems (Hive, Impala, HDFS). · Solid experience in data warehousing and data modelling techniques. More ❯
RELOCATION ASSISTANCE: No relocation assistance available CLEARANCE TYPE: Secret TRAVEL: Yes, 10% of the Time Description At Northrop Grumman, our employees have incredible opportunities to work on revolutionary systems that impact people's lives around the world today, and for More ❯
London, England, United Kingdom Hybrid / WFH Options
Axiom Software Solutions Limited
be a bonus) SQL PySpark Delta Lake Bash (both CLI usage and scripting) Git Markdown Scala (bonus, not compulsory) Azure SQL Server as a HIVE Metastore (bonus) Technologies: Azure Databricks Apache Spark Delta Tables Data processing with Python PowerBI (Integration/Data Ingestion) JIRA #J-18808-Ljbffr More ❯
be a bonus) SQL PySpark Delta Lake Bash (both CLI usage and scripting) Git Markdown Scala (bonus, not compulsory) Azure SQL Server as a HIVE Metastore (bonus) Technologies: Azure Databricks Apache Spark Delta Tables Data processing with Python PowerBI (Integration/Data Ingestion) JIRA Due to the nature More ❯
of Data Mining, Classical Machine Learning, Deep Learning, NLP and Computer Vision. Experience with Large Scale/Big Data technology, such as Hadoop, Spark, Hive, Impala, PrestoDb. Hands-on capability developing ML models using open-source frameworks in Python and R and applying them on real client use cases. … Proficient in one of the deep learning stacks such as PyTorch or Tensorflow. Working knowledge of parallelisation and async paradigms in Python, Spark, Dask, Apache Ray. An awareness and interest in economic, financial and general business concepts and terminology. Excellent written and verbal command of English. Strong problem-solving More ❯
in large-scale data migration efforts. (Mandatory) Demonstrated experience with database architecture, performance design methodologies, and system-tuning recommendations. Preference for familiarity with Glue, Hive, and Iceberg or similar Demonstrated experience with Python, Bash, and Terraform Demonstrated experience with DevSecOps solutions and tools Demonstrated experience implementing CI/CD … and Data Governance concepts and experience. Demonstrated experience maintaining, supporting, and improving the ETL process through the implementation and standardization of data flows with Apache Nifi and other ETL tools. Demonstrated experience with Apache Spark More ❯
large-scale data migration efforts. 5. (Mandatory) Demonstrated experience with database architecture, performance design methodologies, and system-tuning recommendations. Preference for familiarity with Glue, Hive, and Iceberg or similar 6. (Mandatory) Demonstrated experience with Python, Bash, and Terraform 7. (Mandatory) Demonstrated experience with DevSecOps solutions and tools 8. (Mandatory … Governance concepts and experience. 11. (Desired) Demonstrated experience maintaining, supporting, and improving the ETL process through the implementation and standardization of data flows with Apache Nifi and other ETL tools. 12. (Desired) Demonstrated experience with Apache Spark More ❯
London, England, United Kingdom Hybrid / WFH Options
Timely Find
languages and carrying out data analysis and hypothesis testing - Advanced SQL OR Python. Experience with big data technologies and data platforms - we use BigQuery, Apache Ibis, SQLGlot, DBT. You might have experience with Hadoop, Hive, Redshift, Snowflake, Spark or similar. Experience with Version control/CI/CD More ❯
London, England, United Kingdom Hybrid / WFH Options
Autodesk
code, architectures, and experiments Relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra) Frameworks such as Ray data, Metaflow, Hadoop, Spark, or Hive Preferred Qualifications Experience with computational geometry such as mesh or boundary representation data processing. Experience with CAD model search and retrieval, in PLM systems … code, architectures, and experiments Relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra) Frameworks such as Ray data, Metaflow, Hadoop, Spark, or Hive Vector data stores Preferred Qualifications Experience with computational geometry such as mesh or boundary representation data processing. Experience with CAD model search and retrieval More ❯
Business Research Analyst - II, RBS Returns Reduction As a Research Analyst, you'll collaborate with experts to develop ML models leveraging big data solutions and Large Language Models (LLMs) for business needs. You'll drive product pilots, demonstrating innovative thinking More ❯