dynamic data models, ensuring adaptability based on evolving datasets. - Implement caching and filtering techniques to improve query performance. - Utilize big data technologies such as ApacheSpark for large-scale data processing. - Apply AWS expertise, including Infrastructure-as-Code (IaC) for scalable deployment. - Work with SQL transpilers and predicate More ❯
dynamic data models, ensuring adaptability based on evolving datasets. - Implement caching and filtering techniques to improve query performance. - Utilize big data technologies such as ApacheSpark for large-scale data processing. - Apply AWS expertise, including Infrastructure-as-Code (IaC) for scalable deployment. - Work with SQL transpilers and predicate More ❯
dynamic data models, ensuring adaptability based on evolving datasets. - Implement caching and filtering techniques to improve query performance. - Utilize big data technologies such as ApacheSpark for large-scale data processing. - Apply AWS expertise, including Infrastructure-as-Code (IaC) for scalable deployment. - Work with SQL transpilers and predicate More ❯
Strong skills in Python and Pandas, with experience in converting and optimising CPU-based models and algorithms to run efficiently on GPUs. Familiarity with Spark and/or PySpark for handling large-scale data processing tasks. Deep understanding of machine learning techniques and approaches, ensuring best practices in model More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citigroup Inc
Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other technology counterparts … escalating, managing and reporting control issues with transparency. Skills & Qualifications: Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage, capacity planning, investigating. Perform daily More ❯
Belfast. Your day-to-day -Wrangle and draw meaningful insights from massive amounts of unstructured textual data using the latest tools and technologies like Spark, Iceberg, Athena, AWS SageMaker -Apply unsupervised learning algorithms across billions of email interactions to identify emerging threat patterns -Work with state-of-the-art More ❯
deployment pipelines Proficient in cloud AI services (AWS SageMaker/Bedrock) Deep understanding of distributed systems and microservices architecture Expert in data pipeline platforms (Apache Kafka, Airflow, Spark) Proficient in both SQL (PostgreSQL, MySQL) and NoSQL (Elasticsearch, MongoDB) databases Strong containerization and orchestration skills (Docker, Kubernetes) Experience with More ❯
deployment pipelines Proficient in cloud AI services (AWS SageMaker/Bedrock) Deep understanding of distributed systems and microservices architecture Expert in data pipeline platforms (Apache Kafka, Airflow, Spark) Proficient in both SQL (PostgreSQL, MySQL) and NoSQL (Elasticsearch, MongoDB) databases Strong containerization and orchestration skills (Docker, Kubernetes) Experience with More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citigroup Inc
experience in an Application Development role. Demonstrated execution capabilities. Strong analytical and quantitative skills; Data driven and results-oriented Experience with Core Java required (Spark a plus) Experience with SQL Experience working with Hadoop, Hive, Sqoop and other technologies in Cloudera's CDP distribution. Understanding of version control (git … working as part of an agile team. Excellent written and oral communication skills Technical Skills: Strong knowledge in Java Some knowledge inHadoop, hive, SQL, Spark Understanding of Unix Shell Scripting CI/CD Pipeline Maven or Gradle experience Predictive analytics (desirable) PySpark (desirable) Trade Surveillance domain knowledge (desirable) Education More ❯
Belfast, County Antrim, Northern Ireland, United Kingdom
Experis
Role: Data Engineer Location: Belfast Duration: Long Term Contract Opportunity - Rolling Contract Rate: Market Rates - Inside IR35 Job Description : Fujitsu's Decision Intelligence practice in the UK helps organisations bridge the gap between data and insights, empowering businesses to make More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citigroup Inc
Reliable, Relevant, Consistent, Complete, Scalable, Timely, Secure, Nimble. Olympus is built on Big data platform and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Our team interfaces with a vast client base and works in close partnership with Operations, Development and other technology counterparts … role. Hands-on experience in supporting applications built in Hadoop. Working knowledge of various components and technologies under Cloudera distribution like HDFS, Hive, Impala, Spark, YARN, Sentry, Oozie, Kafka. Experienced in Linux Very good knowledge on analyzing the bottlenecks on the cluster - performance tuning, effective resource usage, capacity planning More ❯
Systems Architecture - AWS data engineer Job Description Location: UK - Will require travel to customer site ](Belfast) Job Summary: We are seeking a skilled and experienced AWS Data Engineer to join our team. The successful candidate will be responsible for implementing More ❯
of cloud-based AI tools (e.g. AWS SageMaker, Bedrock) Experienced in building retrieval-augmented generation (RAG) systems Proficient in streaming and orchestration tools like Apache Kafka, Airflow, Spark, Docker, and Kubernetes Solid knowledge of both SQL and NoSQL databases Background in distributed systems and microservices Committed to ethical More ❯
of cloud-based AI tools (e.g. AWS SageMaker, Bedrock) Experienced in building retrieval-augmented generation (RAG) systems Proficient in streaming and orchestration tools like Apache Kafka, Airflow, Spark, Docker, and Kubernetes Solid knowledge of both SQL and NoSQL databases Background in distributed systems and microservices Committed to ethical More ❯
Archival & Reporting application. Core Java, Spring and Hibernate will be the primary technologies used supplemented by strong technical knowledge of Oracle databases Knowledge of Spark, Big Data is desirable but not mandatory. The candidate should have hands on experience in designing and developing solutions The candidate will apply internal … and mockito. Experience in any cloud technologies like Openshift/PCF/AWS/GCP. Experience with building distributed systems, using solutions such as Spark, Big Data Technologies would be preferred but not mandatory. Knowledge of Big Data querying tools (Cloudera stack or similar) e.g. Hive or Impala would More ❯