accountable. These are ourvalues and influence everything we do. Preferred Qualifications Experience with PL/SQL or other database procedural language Experience with Hadoop technologies such as HDFS, Pig, Hive, Impala and management of the Hadoop system. Experience with Azure, AWS, or OCI including their native database services. Experience with Snowflake Experience with Kafka and Elasticsearch Experience with SAS More ❯
development/product management environmenta Relevant experience within core java and spark Experience in systems analysis and programming of java applications Experience using big data technologies (e.g. Java Spark, hive, Hadoop) Ability to manage multiple/competing priorities and manage deadlines or unexpected changes in expectations or requirements Prior financial services/trade surveillance experience is desirable Strong analytical More ❯
such as bash, Python, or Go Must have a DoD 8140/8570 compliance certification (i.e. Security+ certification) Preferred Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes More ❯
GPU workloads, and distributed storage, including Cloudera Experience in the development of algorithms leveraging R, Python, SQL, or NoSQL Experience with Distributed data or computing tools, including MapReduce, Hadoop, Hive, EMR, Spark, Gurobi, or MySQL Experience with visualization packages, including Plotly, Seaborn, or ggplot2 More ❯
indexing/search, and GPU workloads Experience in the development of algorithms leveraging R, Python, or SQL/NoSQL Experience with Distributed data/computing tools, including MapReduce, Hadoop, Hive, EMR, Spark, Gurobi, or MySQL Experience with visualization packages, including Plotly, Seaborn, or ggplot2 Bachelor's degree More ❯
R, Matlab, SAS Enterprise Miner Elastic search and understanding of Hadoop ecosystem Experience working with large data sets, experience working with distributed computing tools like Map/Reduce, Hadoop, Hive, Pig etc. Advanced use of Excel spread sheets for analytical purposes An MSc or PhD in Data Science or an analytical subject (Physics, Mathematics, Computing) or other quantitative discipline More ❯
expertise in Hadoop, Spark, and related technologies YOUR PROFILE Expertise on Hadoop, Spark & Scala Experience in developing complex data transformation workflows(ETL) using Big Data Technologies Good expertise on HIVE, Impala, HBase Hands on experience to finetune Spark jobs Experience with Java and distributed computing ABOUT CAPGEMINI Capgemini is a global business and technology transformation partner, helping organizations to More ❯
normalization, etc. Proficiency in Java/Spring web application development. Experience with Test Driven Development and Agile methodologies; Behavior Driven Development is a plus. Knowledge of Hadoop, Big Data, Hive, Pig, NoSQL is a plus, though most engineers with this background may have limited REST experience. Additional Information All your information will be kept confidential according to EEO guidelines. More ❯
SCI with Full Scope Polygraph. Desired Qualifications: Experience with AWS Data Management services (Elastic Map Reduce, Lambda, Kinesis). Experience with SAFe development practices. Experience with Python, SpringBoot, Hibernate, Hive, Pig, or C++. More ❯
tools to automate profit-and-loss forecasting and planning for the Physical Consumer business. We are building the next generation Business Intelligence solutions using big data technologies such as Apache Spark, Hive/Hadoop, and distributed query engines. As a Data Engineer in Amazon, you will be working in a large, extremely complex and dynamic data environment. You More ❯
Big Data Store (NoSQL) such as Hbase, CloudBase/Acumulo, Big Table, etc.; Shall have demonstrated work experience with the Map Reduce programming model and technologies such as Hadoop, Hive, Pig, etc.; Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS); Shall have demonstrated work experience with Serialization such as JSON and/or BSON More ❯
Computer Science, Statistics, Applied Mathematics, or Engineering - Strong experience with Python and R - A strong understanding of a number of the tools across the Hadoop ecosystem such as Spark, Hive, Impala & Pig - An expertise in at least one specific data science area such as text mining, recommender systems, pattern recognition or regression models - Previous experience in leading a team More ❯
Data Scientist - skills in statistics, physics, mathematics, Computer Science, Engineering, Data Mining, Big Data (Hadoop, Hive, MapReduce) This is an exceptional opportunity to work as a Data Scientist within a global analytics team, utilizing various big data technologies to develop complex behavioral models, analyze customer uptake of products, and foster new product innovation. Responsibilities include: Generating and reviewing large More ❯
Naive Bayes, Random Forests, etc. - Experience with common data science toolkits, such as Python - Proficiency in using query languages such as SQL on a big data platform e.g. Hadoop, Hive - Good applied statistics skills, such as distributions, statistical testing, regression, etc. - Good scripting and programming skills It would be desirable for the successful candidate to come from a Media More ❯
performance management tools. Experience with one or more of the following data tools: Tableau, Foresight, GCP or SQL The other stuff we are looking for Shell scripting, Python, Spark, HIVE, NiFi, Hortonworks/Cloudera DataFlow, HDFS. What's in it for you Our goal is to celebrate our people, their lives and everything in-between. We aim to create More ❯
performance management tools. Experience with one or more of the following data tools: Tableau, Foresight, GCP or SQL The other stuff we are looking for Shell scripting, Python, Spark, HIVE, NiFi, Hortonworks/Cloudera DataFlow, HDFS. What's in it for you Our goal is to celebrate our people, their lives and everything in-between. We aim to create More ❯
performance management tools. Experience with one or more of the following data tools: Tableau, Foresight, GCP or SQL The other stuff we are looking for Shell scripting, Python, Spark, HIVE, NiFi, Hortonworks/Cloudera DataFlow, HDFS. What's in it for you Our goal is to celebrate our people, their lives and everything in-between. We aim to create More ❯
performance management tools. Experience with one or more of the following data tools: Tableau, Foresight, GCP or SQL The other stuff we are looking for Shell scripting, Python, Spark, HIVE, NiFi, Hortonworks/Cloudera DataFlow, HDFS. What's in it for you Our goal is to celebrate our people, their lives and everything in-between. We aim to create More ❯
Bristol, Avon, South West, United Kingdom Hybrid / WFH Options
Hargreaves Lansdown
JIRA. Expertise in implementing statistical techniques in production-grade code, with a focus on scalability and reliability. Experience with large-scale data analysis, manipulation, and distributed computing platforms (e.g., Hive, Hadoop). Familiarity with advanced machine learning methods, including neural networks, reinforcement learning, and other cutting-edge Gen AI approaches. Skilled in API development and deployment, with a focus More ❯
Employment Type: Permanent, Part Time, Work From Home
networks into production Experience with Docker Experience with NLP and/or computer vision Exposure to cloud technologies (eg. AWS and Azure) Exposure to Big data technologies Exposure to Apache products eg. Hive, Spark, Hadoop, NiFi Programming experience in other languages This is not an exhaustive list, and we are keen to hear from you even if you More ❯
work experience with: o Distributed scalable Big Data Store (NoSQL) such as Hbase, CloudBase/Acumulo, Big Table, etc. o Map Reduce programming model and technologies such as Hadoop, Hive, Pig, etc. o Hadoop Distributed File System (HDFS) o Serialization such as JSON and/or BSON • 4 years of SWE experience may be substituted for a bachelor's More ❯
The role also involves optimizing database architecture and performance, implementing DevSecOps practices, and building CI/CD pipelines using Python, Bash, and Terraform. Preferred candidates will have experience with Apache Spark, Apache Nifi, data governance, and ETL standardization. Familiarity with Glue, Hive, and Iceberg or similar technologies is a plus. Tasks Performed: • Bridge communication between technical staff … data between systems, and optimize queries. • Plan and execute large-scale data migrations. • Improve database performance through architecture and tuning. • Create and maintain data flows using ETL tools like Apache Nifi. • Manage infrastructure as code using Python, Bash, and Terraform. • Integrate security into development and deployment workflows. • Build and support automated CI/CD pipelines. Education, Experience and Qualifications … mapping between SQL databases. • Demonstrated experience in large-scale data migration efforts. • Demonstrated experience with database architecture, performance design methodologies, and system-tuning recommendations. Preference for familiarity with Glue, Hive, and Iceberg or similar. • Demonstrated experience with Python, Bash, and Terraform. • Demonstrated experience with DevSecOps solutions and tools. • Demonstrated experience implementing CI/CD pipelines using industry standard process. More ❯
support highly distributed, massively parallel computation needssuch as Hbase, CloudBase/Acumulo, Big Table. Shall have demonstrated work experience with the Map Reduce programming model and technologiessuch as Hadoop, Hive, Pig. Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS). Shall have demonstrated work experience with Serialization such as JSON and/or BSON. Shall More ❯
and translating complex technical issues for non-technical audiences. Self-motivated and able to work independently. Preferred Qualifications: Background in investment banking or financial services. Hands-on experience with Hive, Impala, and the Spark ecosystem (e.g., HDFS, Apache Spark, Spark-SQL, UDFs, Sqoop). Proven experience building and optimizing big data pipelines, architectures, and data sets. More ❯