Lake, Synapse, Power BI) Required Skills & Experience Proven experience as a Data Architect in enterprise environments Extensive hands-on experience with Databricks (including SQL, PySpark, Delta Lake) Solid background in data warehousing , data lakes , and big data frameworks Strong knowledge of Azure cloud services , especially in data integration Experience More ❯
Bristol, Gloucestershire, United Kingdom Hybrid / WFH Options
Snap Analytics
cloud-based architectures on AWS, Azure, or GCP A strong background in Databricks for building enterprise data warehouses with strong knowledge of SQL and Pyspark A deep understanding of data modelling and advanced pipeline development Knowledge and understanding of best practice CI/CD approaches utilising Git. Strategic Vision More ❯
Tampa, Florida, United States Hybrid / WFH Options
LTIMindtree
Python Unit Test libraries Good to have GenAI skillset Good knowledge and work experience in Unixcomm and Shellscripts etc Good to have experience on Pyspark Hadoop and Hive as well Expertise in software engineering principles such as design patterns code design testing and documentation Writing effective and scalable codes More ❯
libraries, Good to have Gen-AI skillset. Good knowledge and work experience in Unix comm and Shell scripts etc Good to have experience on Pyspark Hadoop and Hive as well Expertise in software engineering principles such as design patterns code design testing and documentation Writing effective and scalable codes More ❯
Python, SQL, and big data technologies (Hadoop, Spark, NoSQL) Hands-on experience with cloud platforms (AWS, GCP, Azure) Proficiency in data processing frameworks like PySpark A problem-solver who thrives in a fast-paced environment Excellent communication skills to collaborate with technical and non-technical stakeholders Salary: Up to More ❯
Python, SQL, and big data technologies (Hadoop, Spark, NoSQL) Hands-on experience with cloud platforms (AWS, GCP, Azure) Proficiency in data processing frameworks like PySpark A problem-solver who thrives in a fast-paced environment Excellent communication skills to collaborate with technical and non-technical stakeholders Salary: Up to More ❯
Python, SQL, and big data technologies (Hadoop, Spark, NoSQL) Hands-on experience with cloud platforms (AWS, GCP, Azure) Proficiency in data processing frameworks like PySpark A problem-solver who thrives in a fast-paced environment Excellent communication skills to collaborate with technical and non-technical stakeholders Salary: Up to More ❯
Python, SQL, and big data technologies (Hadoop, Spark, NoSQL) Hands-on experience with cloud platforms (AWS, GCP, Azure) Proficiency in data processing frameworks like PySpark A problem-solver who thrives in a fast-paced environment Excellent communication skills to collaborate with technical and non-technical stakeholders Salary: Up to More ❯
a modern tech stack including SQL, Python, Airflow, Kubernetes, and various other cutting-edge technologies. You'll work with tools like dbt on Databricks, PySpark, Streamlit, and Django, ensuring robust data infrastructure that powers business-critical operations. What makes this role particularly exciting is the combination of technical depth More ❯
a modern tech stack including SQL, Python, Airflow, Kubernetes, and various other cutting-edge technologies. You'll work with tools like dbt on Databricks, PySpark, Streamlit, and Django, ensuring robust data infrastructure that powers business-critical operations. What makes this role particularly exciting is the combination of technical depth More ❯
experience (at minimum) working with modern relational databases and/or distributed computing platforms Big Data, and their query interfaces, such as SQL, Spark, PySpark and Hive. Academic experience (at minimum) using visualization techniques for presenting data and analysis as dashboard in tools such as R/Shiny, GGPlot More ❯
influence. A drive to learn new technologies and techniques. Experience/aptitude towards research and openness to learn new technologies. Experience with Azure, Spark (PySpark), and Kubeflow - desirable. We pay competitive salaries based on experience of the candidates. Along with this, you will be entitled to an award-winning More ❯
influence. A drive to learn new technologies and techniques. Experience/aptitude towards research and openness to learn new technologies. Experience with Azure, Spark (PySpark), and Kubeflow - desirable. We pay competitive salaries based on experience of the candidates. Along with this, you will be entitled to an award-winning More ❯
languages e.g. Python, R, Scala, etc.; (Python preferred). Proficiency in database technologies e.g. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. pySpark, Hive, etc. Experienced working with structured and unstructured data e.g. Text, PDFs, jpgs, call recordings, video, etc. Knowledge of machine learning modelling techniques and More ❯
languages eg. Python, R, Scala, etc.; (Python preferred). Proficiency in database technologies eg. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. pySpark, Hive, etc. Experienced working with structured and also unstructured data eg. Text, PDFs, jpgs, call recordings, video, etc. Knowledge of machine learning modelling techniques More ❯
coding languages eg. Python, R, Scala, etc.; (Python preferred) Proficiency in database technologies eg. SQL, ETL, No-SQL, DW, and Big Data technologies e.g. pySpark, Hive, etc. Experienced working with structured and also unstructured data eg. Text, PDFs, jpgs, call recordings, video, etc. Knowledge of machine learning modelling techniques More ❯
London, England, United Kingdom Hybrid / WFH Options
trg.recruitment
Rate: Up to £600 per day Contract: 6 months (Outside IR35, potential to go perm) Tech Stack: Azure Data Factory, Synapse, Databricks, Delta Lake, PySpark, Python, SQL, Event Hub, Azure ML, MLflow We’ve partnered with a new AI-first professional services consultancy that’s taking on the Big More ❯
Platform technologies (Synapse, Data Lakes, ADF) Expertise in data modelling, ETL/ELT pipeline development, and data integration Proficient in SQL and Python (ideally PySpark) Knowledge of tools such as Power BI, Microsoft Fabric, and DevOps (CI/CD pipelines) Experience working with enterprise data sources and APIs (e.g. More ❯
s or PhD in a relevant field (e.g., Computer Science, Data Science, Engineering, Applied Mathematics, Statistics, etc.). Proficiency in Python, SQL, AWS, Airflow, PySpark, PyTorch, NumPy, and related data technologies. Experience with cloud infrastructure, data pipelines, and machine learning model deployment. Proven experience leading diverse teams of data More ❯
s or PhD in a relevant field (e.g., Computer Science, Data Science, Engineering, Applied Mathematics, Statistics, etc.). Proficiency in Python, SQL, AWS, Airflow, PySpark, PyTorch, NumPy, and related data technologies. Experience with cloud infrastructure, data pipelines, and machine learning model deployment. Proven experience leading diverse teams of data More ❯
commercial experience is a plus Strong understanding of mathematical background, focusing on statistics and linear algebra Highly proficient in Python (Pandas, Scikit-Learn, PyTorch, PySpark) and SQL Experience with Snowflake (function & procedure) and Snowpark is a plus Experience with unit and integration tests Strong understanding of machine learning algorithms More ❯
positive influence on Mesh-AI, our customers, and your team. Nice to Have NLP, LLMs, GenAI, time series forecasting, image recognition or deep learning. PySpark, OpenCV, spaCy or DVC. Exposure to MLOps. Want to know more? Get in touch with careers@mesh-ai.com . Otherwise, apply here. #J More ❯
Hue, Zookeeper, HCatalog, Solr, Avro, Parquet, Iceberg, Hudi). Experience developing software and data engineering code in one or more programming languages (Java, Python, PySpark, Node, etc). AWS and other Data and AI aligned Certifications. Ability to think strategically about business, product, and technical challenges in an enterprise More ❯
Oozie, Hue, Zookeeper, HCatalog, Solr, Avro, Parquet, Iceberg, Hudi) - Experience developing software and data engineering code in one or more programming languages (Java, Python, PySpark, Node, etc) - AWS and other Data and AI aligned Certifications PREFERRED QUALIFICATIONS - Ability to think strategically about business, product, and technical challenges in an More ❯
the-art data analytics platform on AWS, employing the AWS Cloud Development Kit (CDK). Construct resilient and scalable data pipelines using SQL/PySpark/Airflow to effectively ingest, process, and transform substantial data volumes from diverse sources into a structured format, ensuring data quality and integrity. Devise More ❯