with cloud platforms (AWS, Azure, GCP) and deploying models. Ability to use data visualization tools like Tableau or Power BI. Nice-to-Have: Familiarity with big data tools like Hadoop, Kafka, Spark. Knowledge of data governance and validation standards in energy. Experience with distributed computing and large-scale deployment. Strong communication skills for explaining complex validation results. At GE More ❯
London, England, United Kingdom Hybrid / WFH Options
Simon-Kucher & Partners
stand out: Implementation experience with Machine Learning models and applications Knowledge of cloud-based Machine Learning engines (AWS, Azure, Google, etc.) Experience with large scale data processing tools (Spark, Hadoop, etc.) Ability to query and program databases (SQL, No SQL) Experience with distributed ML frameworks (TensorFlow, PyTorch, etc.) Familiarity with collaborative software tools (Git, Jira, etc.) Experience with user More ❯
TB) data sets PREFERRED QUALIFICATIONS - Master's degree in statistics, data science, or an equivalent quantitative field - Experience using Cloud Storage and Computing technologies such as AWS Redshift, S3, Hadoop, etc. - Experience programming to extract, transform and clean large (multi-TB) data sets - Experience with AWS technologies Amazon is an equal opportunities employer. We believe passionately that employing a More ❯
Great experience as a Data Engineer Experience with Spark, Databricks, or similar data processing tools. Proficiency in working with the cloud environment and various software’s including SQL Server, Hadoop, and NoSQL databases. Proficiency in Python (or similar), SQL and Spark. Proven ability to develop data pipelines (ETL/ELT). Strong inclination to learn and adapt to new More ❯
Great experience as a Data Engineer Experience with Spark, Databricks, or similar data processing tools. Proficiency in working with the cloud environment and various software’s including SQL Server, Hadoop, and NoSQL databases. Proficiency in Python (or similar), SQL and Spark. Proven ability to develop data pipelines (ETL/ELT). Strong inclination to learn and adapt to new More ❯
projects demonstrating your ability to solve problems with innovative ideas or coding challenges. Familiarity with Agile practices in a collaborative team setting. Exposure to big data tools, such as Hadoop and Spark, for managing large-scale datasets. Experience with cloud platforms like Microsoft Azure. Why Join? They have created a super collaborative culture offering a great balance of impact More ❯
curious, and driven to innovate continuously. Ability to manage multiple tasks and projects under tight deadlines. Experience with software development (e.g., Python, Java, Scala) and Big Data tools (e.g., Hadoop, Spark), ideally in a cloud-based environment (AWS, Azure, or GCP). Proficiency in SQL. Experience using SAP SuccessFactors and certification in SuccessFactors modules. An Inclusive Workplace City Football More ❯
in a technology company - Experience in data mining, ETL, etc. and using databases in a business environment with large-scale, complex datasets - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience operating large data warehouses - Bachelor's/Master's degree in computer science, engineering, analytics, mathematics, statistics, IT or equivalent Amazon is an equal opportunities More ❯
London, England, United Kingdom Hybrid / WFH Options
Emperia
and attention to detail. Excellent communication skills to translate data insights into business value Preferred Qualifications Experience with serverless architectures on AWS. Familiarity with big data technologies (e.g. Spark, Hadoop). Knowledge of data governance and compliance best practices. Background in software development or experience with APIs for data integration About You Excellent communication and teamwork skills. Ability to More ❯
statistical analysis. Expertise in Python, with proficiency in ML and NLP libraries such as Scikit-learn, TensorFlow, Faiss, LangChain, Transformers and PyTorch. Experience with big data tools such as Hadoop, Spark, and Hive. Familiarity with CI/CD and MLOps frameworks for building end-to-end ML pipelines. Proven ability to lead and deliver data science projects in an More ❯
London, England, United Kingdom Hybrid / WFH Options
EDB
years of experience in Data Warehouse solution with an emphasis on on Greenplum In depth knowledge of Greenplum’s parallel processing and integration with big data frameworks such as Hadoop Strong understanding of data warehousing concepts and analytical query design Strong understanding of Kubernetes architecture (nodes, pods, services, deployments, etc.), knowledge of Kubernetes API objects and their relationships, proficiency More ❯
Prior experience as a Data Engineer . Experience with Spark, Databricks, or similar data processing tools. Proficiency in working with the cloud environment and various software, including SQL Server, Hadoop, and NoSQL databases. Proficiency in Python (or similar), SQL and Spark. Proven ability to develop data pipelines (ETL/ELT). Strong inclination to learn and adapt to new More ❯
scale data pipelines. Experience in dealing with streaming and batch compute frameworks like Spring Kafka, Kafka Streams, Flink, Spark Streaming, Spark. Experience with large-scale computing platforms such as Hadoop, Hive, Spark, and NoSQL stores. Exposure to UI development of visualisations in modern web apps (e.g. Streamlit, Retool) is nice to have. Experience with AWS, GCP. Benefits Roku is More ❯
Telford, England, United Kingdom Hybrid / WFH Options
hackajob
development environment Experience estimating task effort and identifying dependencies Excellent communication skills Familiarity with Python and its numerical, data and machine learning libraries Favourable If You Have Experience of Hadoop and Jenkins Azure Certified AWS Certified Familiarity with Java This position is a full time, permanent role and applicants must have (or be able to acquire) SC clearance. Ad More ❯
exceed their expectations Sales acumen, identifying and managing sales opportunities at client engagements An understanding of database technologies e.g. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. Hadoop, Mahout, Pig, Hive, etc.; An understanding of statistical modelling techniques e.g. Classification and regression techniques, Neural Networks, Markov chains, etc.; An understanding of cloud technologies e.g. AWS, GCP or More ❯
automated testing and deployments using CI/CD pipelines. Continual learning through internal and external training. What you'll bring Mandatory Proficient in at least two of the following: Hadoop, GCP or AWS for creating Big Data solutions. Skilled in using technologies such as Scala or Java. Hands-on experience with Continuous Integration and Deployment strategies. Solid understanding of More ❯
exceed their expectations Sales acumen, identifying and managing sales opportunities at client engagements An understanding of database technologies e.g. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. Hadoop, Mahout, Pig, Hive, etc.; An understanding of statistical modelling techniques e.g. Classification and regression techniques, Neural Networks, Markov chains, etc.; An understanding of cloud technologies e.g. AWS, GCP or More ❯
Belfast, Northern Ireland, United Kingdom Hybrid / WFH Options
Citigroup Inc
About the Team: Data Solutions Technology strives to provide measurable competitive advantage to our business by delivering high quality, innovative and cost effective reference data technology and operational; solutions in order to meet the needs of the business, our clients More ❯
travel. Requirements: 2+ years in customer-facing roles – pre-sales, solutions architect, etc. 7+ years in building enterprise solutions with Middleware, Big Data tools, and frameworks such as Spark, Hadoop, Cassandra, NoSQL, etc. as well as ETL processing. Previous experience with In-Memory Data Grids/Distributed Systems. Proven experience working with various Databases: Relational and NoSQL. Hands-on More ❯
and re-architect enterprise data systems. Migrate on-prem systems to Google Cloud Platform (GCP), leveraging Dataproc, BigQuery, and other GCP-native tooling. Use technologies such as Apache Spark, Hadoop, and Scala to process large-scale distributed datasets. Contribute to infrastructure automation (CI/CD) and hybrid cloud deployment pipelines using tools such as GitHub Actions, Renovate, CodeQL, and More ❯
through mentorship and constructive feedback Who You Are You have data engineering experience and know how to work with high-volume, heterogeneous data, preferably with distributed systems such as Hadoop, BigTable, Cassandra, GCP, AWS, or Azure You have experience with one or more higher-level Python or Java-based data processing frameworks such as Beam, Dataflow, Crunch, Scalding, Storm More ❯
years of relevant experience in several areas of Data Mining, Classical Machine Learning, Deep Learning, NLP and Computer Vision. Experience with Large Scale/Big Data technology, such as Hadoop, Spark, Hive, Impala, PrestoDb. Hands-on capability developing ML models using open-source frameworks in Python and R and applying them on real client use cases. Proficient in one More ❯
Experience developing full stack greenfield applications from inception to development and deployment. Experience using Amazon Web Services (AWS) - EC2, S3, Redshift. Experience of big data and tools such as Hadoop, Spark, etc. Knowledge and experience of object orientated programming and associated concepts. Knowledge and experience of a wide range of open source and commercial technologies. Good understanding of databases More ❯
Terraform, Jenkins, Bamboo, Concourse etc. Monitoring utilising products such as: Prometheus, Grafana, ELK, filebeat etc. Observability - SRE Big Data solutions (ecosystems) and technologies such as: Apache Spark and the Hadoop Ecosystem Edge technologies e.g. NGINX, HAProxy etc. Excellent knowledge of YAML or similar languages The following Technical Skills & Experience would be desirable for Data Devops Engineer: Jupyter Hub Awareness More ❯
predictive modeling, machine-learning, clustering and classification techniques, and algorithms Fluency in a programming language (Python, C,C++, Java, SQL) Familiarity with Big Data frameworks and visualization tools (Cassandra, Hadoop, Spark, Tableau More ❯