City of London, London, United Kingdom Hybrid / WFH Options
Captur
Required Qualifications SQL (BigQuery and PostgreSQL) proficiency and Python programming skills Experience with Google Cloud Platform Experience with big data warehouse systems (Google BigQuery, ApacheHive, etc) Hands on experience working with machine learning teams; understanding of the core concepts of model evaluation techniques and metrics, and suitable More ❯
Required Qualifications SQL (BigQuery and PostgreSQL) proficiency and Python programming skills Experience with Google Cloud Platform Experience with big data warehouse systems (Google BigQuery, ApacheHive, etc) Hands on experience working with machine learning teams; understanding of the core concepts of model evaluation techniques and metrics, and suitable More ❯
London, England, United Kingdom Hybrid / WFH Options
Captur
Required Qualifications SQL (BigQuery and PostgreSQL) proficiency and Python programming skills Experience with Google Cloud Platform Experience with big data warehouse systems (Google BigQuery, ApacheHive, etc) Hands on experience working with machine learning teams; understanding of the core concepts of model evaluation techniques and metrics, and suitable More ❯
Cambourne, England, United Kingdom Hybrid / WFH Options
Remotestar
Familiarity with distributed computing frameworks such as Spark, KStreams, Kafka. Experience with Kafka and streaming frameworks. Understanding of monolithic vs. microservice architectures. Familiarity with Apache ecosystem including Hadoop modules (HDFS, YARN, HBase, Hive, Spark) and Apache NiFi. Experience with containerization and orchestration tools like Docker and Kubernetes. More ❯
automate data ingestion, transformation, and orchestration across systems. Own data operations infrastructure: Manage and optimise key data infrastructure components within AWS, including Amazon Redshift, Apache Airflow for workflow orchestration and other analytical tools. You will be responsible for ensuring the performance, reliability, and scalability of these systems to meet … AWS data services . Strong proficiency in DataOps methodologies and tools, including experience with CI/CD pipelines, containerized applications , and workflow orchestration using Apache Airflow . Familiar with ETL frameworks, and bonus experience with Big Data processing (Spark, Hive, Trino), and data streaming. Proven track record - You More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citi
are Accurate, Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other … practices, and escalating, managing and reporting control issues with transparency. Skills & Qualifications: Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage, capacity planning, investigating. More ❯
London, England, United Kingdom Hybrid / WFH Options
AlphaSights
automate data ingestion, transformation, and orchestration across systems. Own data operations infrastructure: Manage and optimise key data infrastructure components within AWS, including Amazon Redshift, Apache Airflow for workflow orchestration and other analytical tools. You will be responsible for ensuring the performance, reliability, and scalability of these systems to meet … AWS data services . Strong proficiency in DataOps methodologies and tools, including experience with CI/CD pipelines, containerized applications , and workflow orchestration using Apache Airflow . Familiar with ETL frameworks, and bonus experience with Big Data processing (Spark, Hive, Trino), and data streaming. Proven track record – You More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citi
are Accurate, Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other … Application Support role. Hands-on experience in supporting applications built in Hadoop. Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Experienced in Linux Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage More ❯
London, England, United Kingdom Hybrid / WFH Options
Axiom Software Solutions Limited
be a bonus) SQL PySpark Delta Lake Bash (both CLI usage and scripting) Git Markdown Scala (bonus, not compulsory) Azure SQL Server as a HIVE Metastore (bonus) Technologies: Azure Databricks Apache Spark Delta Tables Data processing with Python PowerBI (Integration/Data Ingestion) JIRA #J-18808-Ljbffr More ❯
London, England, United Kingdom Hybrid / WFH Options
Timely Find
languages and carrying out data analysis and hypothesis testing - Advanced SQL OR Python. Experience with big data technologies and data platforms - we use BigQuery, Apache Ibis, SQLGlot, DBT. You might have experience with Hadoop, Hive, Redshift, Snowflake, Spark or similar. Experience with Version control/CI/CD More ❯
London, England, United Kingdom Hybrid / WFH Options
Autodesk
code, architectures, and experiments Relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra) Frameworks such as Ray data, Metaflow, Hadoop, Spark, or Hive Preferred Qualifications Experience with computational geometry such as mesh or boundary representation data processing. Experience with CAD model search and retrieval, in PLM systems … code, architectures, and experiments Relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra) Frameworks such as Ray data, Metaflow, Hadoop, Spark, or Hive Vector data stores Preferred Qualifications Experience with computational geometry such as mesh or boundary representation data processing. Experience with CAD model search and retrieval More ❯
London, England, United Kingdom Hybrid / WFH Options
Northrop Grumman Corp. (AU)
learning, and AI workloads using tools like Jupyter, Spacy, Transformers, and NLTK. Big Data Platforms: Utilize big data NoSQL engines and platforms such as Hive, Impala, and Elasticsearch for data storage and processing. BI and Visualization: Implement and support business intelligence and visualization tools like Tableau, Kibana, and PowerBI … data science, machine learning, and AI tools such as Jupyter, Spacy, Transformers, and NLTK. Experience with big data NoSQL engines/platforms such as Hive, Impala, and Elasticsearch. Proficiency with business intelligence and visualization tools like Tableau, Kibana, and PowerBI. Excellent communication and collaboration skills. Preferred Experience: Certification in More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Vodafone
as ITRS Geneos, AppDynamics Good Experience with Log Aggregation tools such as ELK, Splunk, Grafana(GEM) is preferred Experience working with Oracle Database, Hadoop, Apache Spark, Hive, Starburst Experience with Middleware solutions such as Tibco EMS, Kafka Good written and verbal communication skills What we can offer you More ❯
and-Spoke), security implementations (IAM, Secret Manager, firewalls, Identity-Aware Proxy), DNS configuration, VPN, and Load Balancing. Data Processing & Transformation: Utilize Hadoop cluster with Hive for querying data, and PySpark for data transformations. Implement job orchestration using Airflow. Core GCP Services Management: Work extensively with services like Google Kubernetes …/CD for automation Deep knowledge of network architectures, security implementations, and management of core GCP services Proficiency in employing data processing tools like Hive, PySpark , and data orchestration tools like Airflow Familiarity with managing and integrating diverse data sources Certified GCP Cloud Architect and Data Engineer Additional Information More ❯
Association of Collegiate Conference and Events Directors-International
member of an agile feature team. Help maintain code quality via code reviews. Skill Requirements Proficiency in administrating Big Data technologies (Hadoop, HDFS, Spark, Hive, Yarn, Oozie, Kafka, HBase, Apache stack). Proficiency in defining highly scalable platform architecture. Knowledge of architectural design patterns, highly optimized, low latency More ❯