data-based insights, collaborating closely with stakeholders. Passionately discover hidden solutions in large datasets to enhance business outcomes. Design, develop, and maintain data processing pipelines using Cloudera technologies, including Apache Hadoop, Apache Spark, ApacheHive, and Python. Collaborate with data engineers and scientists to translate data requirements into technical specifications. Develop and maintain frameworks for efficient … verbal communication skills for effective team collaboration. Eagerness to learn and master new technologies and techniques. Experience with AutoSys is preferred. Experience with distributed data/computing tools: Hadoop, Hive, MySQL, etc. If you're a passionate Cloudera Developer eager to make a difference in the banking industry, we want to hear from you! Apply now to join our More ❯
MySQL, PostgreSQL, or Oracle. Experience with big data technologies such as Hadoop, Spark, or Hive. Familiarity with data warehousing and ETL tools such as Amazon Redshift, Google BigQuery, or Apache Airflow. Proficiency in Python and at least one other programming language such as Java, or Scala. Willingness to mentor more junior members of the team. Strong analytical and problem More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming Hands-on experience with multiple databases like PostgreSQL, Snowflake, Oracle, MS SQL Server, NOSQL (HBase/Cassandra, MongoDB) Experience in cloud data eco More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming Hands-on experience with multiple databases like PostgreSQL, Snowflake, Oracle, MS SQL Server, NOSQL (HBase/Cassandra, MongoDB) Experience in cloud data eco More ❯
governance including GDPR. Bonus Points For Expertise in Data Modelling, schema design, and handling both structured and semi-structured data. Familiarity with distributed systems such as Hadoop, Spark, HDFS, Hive, Databricks. Exposure to AWS Lake Formation and automation of ingestion and transformation layers. Background in delivering solutions for highly regulated industries. Passion for mentoring and enabling data engineering best More ❯
data technology with experience ranging from platform architecture, data management, data architecture and application architecture High Proficiency working with Hadoop platform including Spark/Scala, Kafka, SparkSQL, HBase, Impala, Hive and HDFS in multi-tenant environments Solid base in data technologies like warehousing, ETL, MDM, DQ, BI and analytical tools extensive experience in metadata management and data quality processes More ❯
schema design, and GDPR-compliant solutions Working knowledge of DevOps tools and CI/CD processes Bonus Points For Development experience in Scala or Java Familiarity with Cloudera, Hadoop, HIVE, and Spark ecosystem Understanding of data privacy regulations, including GDPR, and experience working with sensitive data Ability to learn and adapt new technologies quickly to meet business needs Collaborative More ❯
with scripting languages like Python or KornShell. Knowledge of writing and optimizing SQL queries for large-scale, complex datasets. PREFERRED QUALIFICATIONS Experience with big data technologies such as Hadoop, Hive, Spark, EMR. Experience with ETL tools like Informatica, ODI, SSIS, BODI, or DataStage. We promote an inclusive culture that empowers Amazon employees to deliver the best results for our More ❯
Practical experience in coding languages eg. Python, R, Scala, etc. (Python preferred) Strong proficiency in database technologies eg. SQL, ETL, No-SQL, DW, and Big Data technologies eg. pySpark, Hive, etc. Experienced working with structured and unstructured data eg. Text, PDFs, jpgs, call recordings, video, etc. Knowledge of machine learning modelling techniques and how to fine-tune those models More ❯
technologies, Virtualisation Tools, Monitoring utilities, Disaster recovery process/tools Experience in troubleshooting and problem resolution Experience in System Integration Knowledge of the following: Hadoop, Flume, Sqoop, Map Reduce, Hive/Impala, Hbase, Kafka, Spark Streaming Experience of ETL tools incorporating Big Data Shell Scripting, Python Beneficial Skills: Understanding of: LAN, WAN, VPN and SD Networks Hardware and Cabling More ❯
expertise in Hadoop, Spark, and related technologies YOUR PROFILE Expertise on Hadoop, Spark & Scala Experience in developing complex data transformation workflows(ETL) using Big Data Technologies Good expertise on HIVE, Impala, HBase Hands on experience to finetune Spark jobs Experience with Java and distributed computing ABOUT CAPGEMINI Capgemini is a global business and technology transformation partner, helping organizations to More ❯
and drive marketing efficiencies by leveraging approaches that optimize Amazon's systems using cutting edge quantitative techniques. The right candidate needs to be fluid in: Data warehousing and EMR (Hive, Pig, R, Python). Feature extraction, feature engineering and feature selection. Machine learning, causal inference, statistical algorithms and recommenders. Model evaluation, validation and deployment. Experimental design and testing. BASIC More ❯
Experience on starting the front-end buildout from scratch by coordinating across multiple business and technology groups o Experience building complex single-page applications using Abinitio/Hadoop/Hive/Kafka/Oracle and modern MOM technologies o Experienced with Linux/Unix platform o Experience in SCMs like GIT; and tools like JIRA o Familiar with the More ❯
R, Matlab, SAS Enterprise Miner Elastic search and understanding of Hadoop ecosystem Experience working with large data sets, experience working with distributed computing tools like Map/Reduce, Hadoop, Hive, Pig etc. Advanced use of Excel spread sheets for analytical purposes An MSc or PhD in Data Science or an analytical subject (Physics, Mathematics, Computing) or other quantitative discipline More ❯
normalization, etc. Proficiency in Java/Spring web application development. Experience with Test Driven Development and Agile methodologies; Behavior Driven Development is a plus. Knowledge of Hadoop, Big Data, Hive, Pig, NoSQL is a plus, though most engineers with this background may have limited REST experience. Additional Information All your information will be kept confidential according to EEO guidelines. More ❯
Computer Science, Statistics, Applied Mathematics, or Engineering - Strong experience with Python and R - A strong understanding of a number of the tools across the Hadoop ecosystem such as Spark, Hive, Impala & Pig - An expertise in at least one specific data science area such as text mining, recommender systems, pattern recognition or regression models - Previous experience in leading a team More ❯
Data Scientist - skills in statistics, physics, mathematics, Computer Science, Engineering, Data Mining, Big Data (Hadoop, Hive, MapReduce) This is an exceptional opportunity to work as a Data Scientist within a global analytics team, utilizing various big data technologies to develop complex behavioral models, analyze customer uptake of products, and foster new product innovation. Responsibilities include: Generating and reviewing large More ❯
Birmingham, Staffordshire, United Kingdom Hybrid / WFH Options
Experis - ManpowerGroup
data sources. Using machine learning tools to select features, create and optimize classifiers Qualifications: Programming Skills - knowledge of statistical programming languages like python, and database query languages like SQL, Hive/Hadoop, Pig is desirable. Familiarity with Scala and java is an added advantage. Statistics - Good applied statistical skills, including knowledge of statistical tests, distributions, regression, maximum likelihood estimators More ❯
Birmingham, West Midlands, United Kingdom Hybrid / WFH Options
Experis
data sources. Using machine learning tools to select features, create and optimize classifiers Qualifications: Programming Skills - knowledge of statistical programming languages like python, and database query languages like SQL, Hive/Hadoop, Pig is desirable. Familiarity with Scala and java is an added advantage. Statistics - Good applied statistical skills, including knowledge of statistical tests, distributions, regression, maximum likelihood estimators More ❯
and optimize classifiers Required Skills/Experience The ideal candidate will have the following: Programming Skills - knowledge of statistical programming languages like python, and database query languages like SQL, Hive/Hadoop, Pig is desirable. Familiarity with Scala and java is an added advantage. Statistics - Good applied statistical skills, including knowledge of statistical tests, distributions, regression, maximum likelihood estimators More ❯
Bristol, Avon, South West, United Kingdom Hybrid / WFH Options
Hargreaves Lansdown
JIRA. Expertise in implementing statistical techniques in production-grade code, with a focus on scalability and reliability. Experience with large-scale data analysis, manipulation, and distributed computing platforms (e.g., Hive, Hadoop). Familiarity with advanced machine learning methods, including neural networks, reinforcement learning, and other cutting-edge Gen AI approaches. Skilled in API development and deployment, with a focus More ❯
Employment Type: Permanent, Part Time, Work From Home
Podman or Docker Comfortable using Git; an awareness of CI/CD practices and tools such as GitHub Actions or Azure DevOps Nice to have: Experience of working with Apache Spark/Flink/Kafka Familiarity with object storage e.g. AWS S3 Knowledge of containerised development workflows using e.g., VSCode Basic understanding of cloud platforms like AWS or GCP … Experience contributing to open-source or internal code templates, demos, or accelerators Familiarity with data catalog services (e.g. Hive, Polaris, Glue) This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required. What well provide you By joining Citi London, you will not only be part of More ❯
Podman or Docker Comfortable using Git; an awareness of CI/CD practices and tools such as GitHub Actions or Azure DevOps Nice to have: Experience of working with Apache Spark/Flink/Kafka Familiarity with object storage e.g. AWS S3 Knowledge of containerised development workflows using e.g., VSCode Basic understanding of cloud platforms like AWS or GCP … Experience contributing to open-source or internal code templates, demos, or accelerators Familiarity with data catalog services (e.g. Hive, Polaris, Glue) This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required. What we'll provide you By joining Citi London, you will not only be part More ❯
deliver innovative solutions. The team's expertise spans a wide range of technologies, including Java and Python-based MicroServices, AWS/GCP cloud backend systems, Big Data technologies like Hive and Spark, and modern Web applications. With a globally distributed presence across the US, India and Europe, the team thrives on collaboration, bringing together diverse perspectives to solve complex … on business needs, entrepreneurial spirit Excellent verbal and written communication skills BS or MS degree in Computer Science or equivalent Nice to Have Experience in distributed computing frameworks like - Hive/Hadoop, Apache Spark Experience in developing Finance or HR related applications Experience with following cloud services: AWS Elastic Beanstalk, EC2, S3, CloudFront, RDS, DynamoDB, VPC, Elastic Cache … Lambda Working experience with Terraform Experience in creating workflows for Apache Airflow Benefits Roku is committed to offering a diverse range of benefits as part of our compensation package to support our employees and their families. Our comprehensive benefits include global access to mental health and financial wellness support and resources. Local benefits include statutory and voluntary benefits which More ❯
deliver innovative solutions. The team's expertise spans a wide range of technologies, including Java and Python-based MicroServices, AWS/GCP cloud backend systems, Big Data technologies like Hive and Spark, and modern Web applications. With a globally distributed presence across the US, India and Europe, the team thrives on collaboration, bringing together diverse perspectives to solve complex … on business needs, entrepreneurial spirit Excellent verbal and written communication skills BS or MS degree in Computer Science or equivalent Nice to Have Experience in distributed computing frameworks like - Hive/Hadoop, Apache Spark Experience in developing Finance or HR related applications Experience with following cloud services: AWS Elastic Beanstalk, EC2, S3, CloudFront, RDS, DynamoDB, VPC, Elastic Cache … Lambda Working experience with Terraform Experience in creating workflows for Apache Airflow Benefits Roku is committed to offering a diverse range of benefits as part of our compensation package to support our employees and their families. Our comprehensive benefits include global access to mental health and financial wellness support and resources. Local benefits include statutory and voluntary benefits which More ❯