knowledge of applying Data Engineering best practices (coding practices to DS, unit testing, version control, code review). * Big Data Eco-Systems, Cloudera/Hortonworks, AWS EMR, GCP DataProc or GCP Cloud Data Fusion. * NoSQL Databases. Dynamo DB/Neo4j/Elastic, Google Cloud Datastore. * Snowflake Data Warehouse/Platform more »
data lakes and data pipelines in cloud using Azure and Databricks or similar tools. Spark Developer certification from any of (Databricks, MAPR, Cloudera or Hortonworks) is added advantage but not required Practice with Unix command line tools Familiarity working with agile methodology Strong database and data analysis skills A history more »
like Tableau Master data management (MDM) – Concepts and expertise in tools like Informatica & Talend MDM Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming Hands-on experience with multiple databases like PostgreSQL, Snowflake, Oracle, MS SQL Server, NOSQL (HBase/ more »
identify key issues, patterns or deviations from norm Skill in identifying data issues and anomalies during analysis Big Data Experience a plus - working with Hortonworks/Hadoop Education and Experience Minimum: BA/BS degree in marketing, economics, statistics, psychology, engineering, computer science, mathematics, or finance Preferred: MA/MS more »