technology stack compatible with AWS. Experience with web scraping and other data ingestion methods and tools. Knowledge of distributed computing frameworks (Hadoop, Spark, Hive, Presto). Experience with data orchestration tools (Airflow, Orchestra, Azkaban). Expertise in cloud data warehousing and core data modelling concepts. Proficiency in version control more »
a cloud-native, modern data engineering technology stack. Experience with data ingestion methods and tools. Experience with distributed computing frameworks (e.g., Hadoop, Spark, Hive, Presto). Experience with data orchestration tools. Experience with cloud data warehousing and core data modelling concepts. Proficiency in version control systems (e.g., Git) and more »
using Terraform, Ansible, or other tools to automate Infrastructure-as-Code that is testable and maintainable Expert in services such as Kafka, Spark, Airflow, Presto, Influx/Cassandra/Dynamo, Microservices, and other technologies used to build data pipelines Experience in developing software projects using Agile/Scrum methodologies more »
eg NGINX, HAProxy etc. Excellent knowledge of YAML or similar languages Desirable Requirements: Jupyter Hub Awareness Minio or similar S3 storage technology Trino/Presto RabbitMQ or other common queue technology eg ActiveMQ NiFi Rego Familiarity with code development, Shell-Scripting in Python, Bash etc. To apply for this more »