Watford, Hertfordshire, United Kingdom Hybrid / WFH Options
Digital Gaming Corp
data from sources like Facebook, Google Analytics, and payment providers. Using tools like AWS Redshift, S3, and Kafka, you'll optimize data models for batch and real-time processing. Collaborating with stakeholders, you'll deliver actionable insights on player behavior and gaming analytics, enhancing experiences and driving revenue with … robust ETL pipelines to integrate data from diverse sources, including APIs like Facebook, Google Analytics, and payment providers. Develop and optimize data models for batchprocessing and real-time streaming using tools like AWS Redshift, S3, and Kafka. Lead efforts in acquiring, storing, processing, and provisioning data More ❯
Manchester, Lancashire, United Kingdom Hybrid / WFH Options
Smart DCC
the data environment. What will you be doing? Design and implement efficient ETL processes for data extraction, transformation, and loading. Build real-time data processing pipelines using platforms like Apache Kafka or cloud-native tools. Optimize batchprocessing workflows with tools like Apache Spark and Flink for … What are we looking for? Advanced proficiency with databases (SQL Server, Oracle, MySQL, PostgreSQL). Expertise in building and managing data pipelines and data processing workflows. Strong understanding of data warehousing concepts, schema design, and data modelling. Hands-on experience with cloud platforms (AWS, Azure, Google Cloud) for scalable More ❯
london, south east england, United Kingdom Hybrid / WFH Options
Intec Select
robust ETL pipelines to integrate data from diverse sources, including APIs like Facebook, Google Analytics, and payment providers. Develop and optimize data models for batchprocessing and real-time streaming using tools like AWS Redshift, S3, and Kafka. Lead efforts in acquiring, storing, processing, and provisioning data More ❯
chunks and deliver them with high quality. Key Responsibilities Distributed Systems Development : Design and build scalable distributed systems using Java-based microservices and Python batchprocessing to support our ML models, evaluation, and observability. Model Lifecycle : Create and maintain robust model deployment pipelines using PySpark and Databricks, ensuring … team members. Complementary skills for this role Technical Expertise : Extensive experience with distributed systems engineering, including designing and implementing Java-based microservices and Python batch jobs. Observability Knowledge : Deep understanding of observability principles, including monitoring, logging, and real-time system insights Data Engineering Skills : Proficiency in building data pipelines More ❯
City, Edinburgh, United Kingdom Hybrid / WFH Options
ENGINEERINGUK
and several other benefits. The individual selected for this position will have the responsibility to cover business-critical compute workloads, real-time/interact processing, data transfer services, application and new technology onboarding, upgrades, and recovery procedures. The international team is split into 4 global regions to provide … implementation and weekend checkouts; aid in incident management and root cause analysis. Provide ongoing operational support for the Aladdin infrastructure. Support and fix both batchprocessing and interactive user applications to ensure the high availability of the Aladdin Environment. Use various tools to conduct analysis on system performance More ❯
the guidelines as an applicant for this role is essential, please read the below carefully. Key Responsibilities: Workflow Design & Development: Design, develop, and optimize batch workflows using Control-M to meet business requirements. System Administration: Install, configure, and maintain Control-M components, including Servers, Agents, and Enterprise Manager. Integration … Integrate Control-M with various applications and databases, ensuring seamless data flow and process automation. Monitoring & Troubleshooting: Monitor batch jobs, troubleshoot failures, and implement solutions to ensure timely execution. Upgrades & Maintenance: Plan and execute Control-M upgrades, patches, and migrations with minimal disruption. Collaboration: Work closely with cross-functional … for automation tasks. Experience with integrating Control-M into cloud environments (AWS, Azure, or GCP). Strong understanding of workload automation, job scheduling, and batchprocessing concepts. Familiarity with observability tools like AppDynamics, Splunk, or Grafana is a plus. Excellent problem-solving skills and the ability to work More ❯
Newcastle Upon Tyne, Tyne And Wear, United Kingdom Hybrid / WFH Options
Accenture
Knowledge of database technologies such as PostgreSQL or relational DB. Knowledge in microservices development and various integration patterns. Knowledge in event driven development and batchprocessing Experience in application development across full stack technologies including integrations with Power Platform, Messaging services, Rabbit MQ, API, Kafka live streaming or More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Adepta Partners
use of associated tools and applications to complete these tasks. Ability to travel to client site, where required, will be a consideration. Experience in processing large amounts of structured and unstructured data, including integrating data from multiple sources through ingestion and curation functions on AWS cloud using AWS native … or custom programming. Knowledge of data mining, machine learning, natural language processing is an advantage. You enjoy working within cross-functional Agile teams and you are familiar with Scrum ceremonies. You’ll be comfortable designing and building for the AWS cloud and will have designed and worked on architectures … AWS best practices. Your AWS experience spans data engineering, data science and product development projects, plus you will have an understanding of stream and batch processing. Sound interesting? Apply now More ❯
belfast, antrim, United Kingdom Hybrid / WFH Options
Anson McCade
analytics services (or equivalent open-source technologies). Expertise in designing and building data lakes, data warehouses, and ETL pipelines. Strong understanding of data processing techniques, including stream and batch processing. Familiarity with data mining, machine learning, and natural language processing is a plus. Ability to travel More ❯