RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & SparkstreamingMore ❯
RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & SparkstreamingMore ❯
and data pipelines. Solid understanding of SQL and relational databases (e.g., MySQL, PostgreSQL, Hive). Familiarity with the Apache Hadoop ecosystem (HDFS, MapReduce, YARN). Working knowledge of ApacheSpark and its modules (e.g., Spark SQL, SparkStreaming, MLlib). Experience with cloud-based data platforms like AWS Glue, Google Cloud Dataflow More ❯
delivering data and reporting solutions that fuel the business. What You'll Do: Be a part of team designing and building Enterprise Level scalable, low-latency, fault-tolerant streaming data platform that provides meaningful and timely insights Build the next generation Distributed Streaming Data Pipelines and Analytics Data Stores using streaming frameworks (Flink, SparkStreaming) using programming languages like Java, Scala, Python Be part of a group of engineers building data pipelines using big data technologies (Spark, Flink, Kafka, Snowflake, AWS Big Data Services, Snowflake, Redshift) on medium to large scale datasets Work in a creative & collaborative environment driven by agile methodologies with focus on CI/CD … including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data computing tools (Kafka, Spark, Flink) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (DynamoDB, OpenSearch) 2+ years of data warehousing experience More ❯
delivering data and reporting solutions that fuel the business. What You'll Do: Be a part of team designing and building Enterprise Level scalable, low-latency, fault-tolerant streaming data platform that provides meaningful and timely insights Build the next generation Distributed Streaming Data Pipelines and Analytics Data Stores using streaming frameworks (Flink, SparkStreaming) using programming languages like Java, Scala, Python Be part of a group of engineers building data pipelines using big data technologies (Spark, Flink, Kafka, Snowflake, AWS Big Data Services, Snowflake, Redshift) on medium to large scale datasets Work in a creative & collaborative environment driven by agile methodologies with focus on CI/CD … including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data computing tools (Kafka, Spark, Flink) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (DynamoDB, OpenSearch) 2+ years of data warehousing experience More ❯
delivering data and reporting solutions that fuel the business. What You'll Do: Be a part of team designing and building Enterprise Level scalable, low-latency, fault-tolerant streaming data platform that provides meaningful and timely insights Build the next generation Distributed Streaming Data Pipelines and Analytics Data Stores using streaming frameworks (Flink, SparkStreaming) using programming languages like Java, Scala, Python Be part of a group of engineers building data pipelines using big data technologies (Spark, Flink, Kafka, Snowflake, AWS Big Data Services, Snowflake, Redshift) on medium to large scale datasets Work in a creative & collaborative environment driven by agile methodologies with focus on CI/CD … including Python, SQL, Scala, or Java 2+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 3+ years experience with Distributed data computing tools (Kafka, Spark, Flink) 2+ year experience working on real-time data and streaming applications 2+ years of experience with NoSQL implementation (DynamoDB, OpenSearch) 2+ years of data warehousing experience More ❯
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: Scala/Spark • Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL • Linux Based … Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) • Experience in Big data technologies, real time data processing platform (SparkStreaming) experience would be an advantage. • Consistently demonstrates clear and concise written and verbal communication • A history of delivering against agreed objectives • Ability to multi-task and work under pressure • Demonstrated problem solving and decision-making skills • Excellent More ❯
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: Scala/Spark • Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL • Linux Based … Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) • Experience in Big data technologies, real time data processing platform (SparkStreaming) experience would be an advantage. • Consistently demonstrates clear and concise written and verbal communication • A history of delivering against agreed objectives • Ability to multi-task and work under pressure • Demonstrated problem solving and decision-making skills • Excellent More ❯
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: Scala/Spark • Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL • Linux Based … Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) • Experience in Big data technologies, real time data processing platform (SparkStreaming) experience would be an advantage. • Consistently demonstrates clear and concise written and verbal communication • A history of delivering against agreed objectives • Ability to multi-task and work under pressure • Demonstrated problem solving and decision-making skills • Excellent More ❯
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: Scala/Spark • Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL • Linux Based … Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) • Experience in Big data technologies, real time data processing platform (SparkStreaming) experience would be an advantage. • Consistently demonstrates clear and concise written and verbal communication • A history of delivering against agreed objectives • Ability to multi-task and work under pressure • Demonstrated problem solving and decision-making skills • Excellent More ❯
london (city of london), south east england, united kingdom
Ubique Systems
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: Scala/Spark • Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL • Linux Based … Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) • Experience in Big data technologies, real time data processing platform (SparkStreaming) experience would be an advantage. • Consistently demonstrates clear and concise written and verbal communication • A history of delivering against agreed objectives • Ability to multi-task and work under pressure • Demonstrated problem solving and decision-making skills • Excellent More ❯
experience across AWS Glue, Lambda, Step Functions, RDS, Redshift, and Boto3. Proficient in one of Python, Scala or Java, with strong experience in Big Data technologies such as: Spark, Hadoop etc. Practical knowledge of building Real Time event streaming pipelines (eg, Kafka, SparkStreaming, Kinesis). Proven experience developing modern data architectures … data governance including GDPR. Bonus Points For Expertise in Data Modelling, schema design, and handling both structured and semi-structured data. Familiarity with distributed systems such as Hadoop, Spark, HDFS, Hive, Databricks. Exposure to AWS Lake Formation and automation of ingestion and transformation layers. Background in delivering solutions for highly regulated industries. Passion for mentoring and enabling data More ❯
relevant to the product being deployed and/or maintained. 5-7 years direct experience in Data Engineering with experience in tools such as: Big data tools: Hadoop, Spark, Kafka, etc. Relational SQL and NoSQL databases, including Postgres and Cassandra. Data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc. AWS cloud services: EC2, EMR, RDS, Redshift (or … Azure equivalents) Data streaming systems: Storm, Spark-Streaming, etc. Search tools: Solr, Lucene, Elasticsearch Object-oriented/object function scripting languages: Python, Java, C++, Scala, etc. Advanced working SQL knowledge and experience working with relational databases, query authoring and optimization (SQL) as well as working familiarity with a variety of databases. Experience with message More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or SparkStreaming) 4+ years of experience in open More ❯
meet functional/non-functional project requirements Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL, streaming and 'big data' technologies Implement data pipelines to ingest data to the platform, standardize and transform the data Support the development of analytics tools that utilize the data pipeline … industry leader. Work with data and analytics experts to strive for greater functionality in our data systems. Design and architect solutions with Big Data technologies (e.g Hadoop, Hive, Spark, Kafka) Design and implement systems that run at scale leveraging containerized deployments Design, build, and scale data pipelines across a variety of source systems and streams (internal, third-party … Science, Computer Engineering, Informatics, Information Systems, or another quantitative field Minimum 5 years of experience in a Data Engineer role Required Skills: Experience with big data tools: Hadoop, Spark, etc. Experience with relational SQL and NoSQL databases, including Postgres Experience with AWS cloud or remote services: EC2, EMR, RDS, Redshift Experience with stream-processing systems: Kafka, Storm, SparkMore ❯