Inmon methodologies. Develop and manage cloud-based data solutions using Azure services such as Data Factory, Event Hubs, Data Lake, Synapse, and Azure SQL Server. Create and optimize data processing workflows in Databricks using PySpark and Spark SQL. Ensure ETL coding standards are met, including self … Kimball, Inmon, and hybrid data warehouse design. Extensive experience with Azure data products including Data Factory, Event Hubs, Data Lake, Synapse, and Azure SQL Server. Proficient in developing with Databricks, PySpark, and Spark SQL. Strong understanding of ETL coding standards, including standardized, self-documenting code and more »
Leeds, England, United Kingdom Hybrid / WFH Options
Damia Group
Spark/PySpark Architect - 12 months+ -£Inside IR35- Hybrid working of 3 days on site in Leeds My client are a Global Consultancy who are looking for a number of Spark/PySpark Architects to join them on a Long term programme. As the Spark … upgrade to PySpark Collaboration with multiple customer stakeholders Knowledge of working with Cloud Databases Excellent communication and solution presentation skills. Able to analyse Spark code failures through Spark Plans and make correcting recommendations Able to review PySpark and SparkSQL jobs and … Able to understand Data Frames/Resilient Distributed Data Sets and understand any memory related problems and make corrective recommendations Able to monitor Spark jobs using wider tools such as Grafana to see whether there are Cluster level failures. As a Spark architect, who can demonstrate more »
governance techniques Good understanding of Quality and Information Security principles Experience with Azure, ETL Tools such as ADF and Databricks Advanced Database and SQL skills, alng with SQL, Python, Pyspark, SparkSQL Strong understanding of data model design and implementation principles Data more »
Schema structure & design, Kimball & inmon, and hybrid data warehouse design. Cloud data products such as: Data factory Events Hubs Data Lake Synapse Azure SQL Server Experience developing Databricks and coding with PySpark and Spark SQL. Proficient in ETL coding standards Data encryption techniques and standards Knowledge more »
Data Modeling principles for relational and dimensional data structures Data Lake design principles, Data Virtualization Strong knowledge of data warehouse concepts and T-SQL relational/non-relational databases for data access and Advanced Analytics Experience with following languages: Python, R, Scala, SQL, M Experience in … multidimensional and/or tabular models (SSAS) Microsoft Azure Stack RDBMS: Azure SQL/SQL Server 2016+ (SQL, Structure, Stored Procs, Tuning) ETL/ELT, (ADF, SSIS) Data governance (Purview, Unity Catalogue) Databricks Delta Lake Storage Azure Dev OPS DESIRED SKILLS Advanced Analytics Data … Technologies Databricks, Delta Lake, Synapse SparkSQL, Pyspark Azure Data Explorer Logic Apps, Key Vault Semi structured data processing Integration Runtime Coding experience: Python, C#, Java for Data analysis purpose One of the following certifications: Azure AI Engineer Associate Azure Fundamentals AI-900 Azure AI Fundamentals more »
years working with data warehouses, relational databases and query languages 2+ years building data pipelines in databricks using pyspark, scala and/or sparkSQL and ability to work across structured, semi-structured and unstructured data 2+ years data modeling (e.g., data vault, star schema, entity more »
of Kimball & Inmon and hybrid data warehouse design Detailed knowledge using cloud data products like Data Factory, Event Hubs, Data Lake, Synapse, Azure SQL server. Detailed knowledge in developing in Databricks and experience in coding with PySpark. SparkSQL ETL coding standards: ensuring that more »
Leeds, West Yorkshire, Yorkshire, United Kingdom Hybrid / WFH Options
Damia Group Ltd
Spark/PySpark Architect - 12 months+ -£Inside IR35- Hybrid working of 3 days on site in Leeds My client are a Global Consultancy who are looking for a number of Spark/PySpark Architects to join them on a Long term programme. As the Spark … upgrade to PySpark Collaboration with multiple customer stakeholders Knowledge of working with Cloud Databases Excellent communication and solution presentation skills. Able to analyse Spark code failures through Spark Plans and make correcting recommendations Able to review PySpark and SparkSQL jobs and … Able to understand Data Frames/Resilient Distributed Data Sets and understand any memory related problems and make corrective recommendations Able to monitor Spark jobs using wider tools such as Grafana to see whether there are Cluster level failures. As a Spark architect, who can demonstrate more »