Strong in PySpark, Delta Lake, Databricks SQL. Experience with Databricks Workflows, Unity Catalog, and Delta Live Tables. Programming & Full Stack: Python (mandatory), SQL (expert). Exposure to Java/Scala (for Spark jobs). Knowledge of APIs, microservices (FastAPI/Flask), or basic front-end (React/Angular) is a plus. Cloud Platforms: Proficiency with at least one: Azure Databricks More ❯
and NoSQL databases such as MongoDB, Cassandra, and HBase. Distributed Systems: Understanding of distributed systems and distributed computing principles. Programming Languages: Proficiency in programming languages such as Java, Python, Scala, and SQL. Data Modeling: Knowledge of data modeling techniques and tools to design efficient data structures for Big Data systems. Data Processing: Experience with data processing and ETL (Extract, Transform More ❯
Spark). Proven expertise in advanced data modeling, real-time data processing using technologies like Kafka or Kinesis, and event-driven architecture. Expert-level proficiency in programming languages (Python, Scala) and SQL. Hands-on experience with containerization (Docker, Kubernetes) and building CI/CD pipelines for data infrastructure. Strong understanding of identity resolution methodologies and privacy-enhancing technologies. Desired Skills More ❯
data pipelines for performance, efficiency, and cost-effectiveness. Implement data quality checks and validation rules within data pipelines. Data Transformation & Processing: Implement complex data transformations using Spark (PySpark or Scala) and other relevant technologies. Develop and maintain data processing logic for cleaning, enriching, and aggregating data. Ensure data consistency and accuracy throughout the data lifecycle. Azure Databricks Implementation: Work extensively More ❯
data pipelines for performance, efficiency, and cost-effectiveness. Implement data quality checks and validation rules within data pipelines. Data Transformation & Processing: Implement complex data transformations using Spark (PySpark or Scala) and other relevant technologies. Develop and maintain data processing logic for cleaning, enriching, and aggregating data. Ensure data consistency and accuracy throughout the data lifecycle. Azure Databricks Implementation: Work extensively More ❯
What we're looking for: Cloud & big data platforms - 5+ years designing and developing on AWS and Databricks data lake and lakehouse architectures Programming - Expert in SQL , Python/Scala Streaming - Hands on with Kafka and Spark Structured Streaming in production Data modeling & warehousing - Dimensional, data vault , and real time schemas; SQL performance tuning on data warehouses , including Snowflake and More ❯
applications and data pipelines What We're Looking for 12+ years of progressive software/data engineering experience with high proficiency in two of the following: Python, Java or Scala Experience in cloud computing (e.g. applications, infrastructure, storage, data and platforms) Designing and building complex, highly scalable, reliable data pipelines, using the Big Data ecosystem (Spark, Iceberg, Glue Catalog, Kafka More ❯
years of experience as a Data Engineer or in a similar role focused on large-scale data systems. Strong programming skills in Python, SQL, and familiarity with Java/Scala a plus. Hands-on experience with big data frameworks (e.g., Spark, Flink, Hadoop) and workflow orchestration (Airflow, Prefect, Dagster). Proven experience with cloud-based data platforms (AWS, GCP, Azure More ❯
Seattle, Washington, United States Hybrid/Remote Options
Pioneer Square Labs
s degree in Computer Science, Data Engineering, or a related field. 5+ years of experience in data engineering, ML engineering, or backend systems. Strong programming skills in Python, SQL, Scala, Java or any OP languages. Proficiency with cloud platforms (AWS, GCP, or Azure) and their data/AI services. Experience with ML pipelines, including feature stores, model registries, and inference More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Billigence
Data architecture and solution design experience Hands-on experience with modern data tools such as dbt, Fivetran, Matillion, or similar data integration platforms Programming skills in Python, Java, or Scala Relevant cloud certifications (SnowPro, Databricks Certified, AWS/Azure/GCP Data Engineering certifications) Experience with DataOps, CI/CD practices, and infrastructure-as-code Knowledge of data governance, data More ❯
Data architecture and solution design experience Hands-on experience with modern data tools such as dbt, Fivetran, Matillion, or similar data integration platforms Programming skills in Python, Java, or Scala Relevant cloud certifications (SnowPro, Databricks Certified, AWS/Azure/GCP Data Engineering certifications) Experience with DataOps, CI/CD practices, and infrastructure-as-code Knowledge of data governance, data More ❯
analytics use cases, with experience in modern data storage solutions like data warehouses (Redshift, BigQuery, Databricks, Snowflake). Additional programming languages: Proficiency in additional languages such as Java or Scala is a plus. Cloud and infrastructure expertise: Experience with AWS cloud platforms and hands on skills in infrastructure as code (SDK, CDK, Terraform) and container orchestration (Docker/Kubernetes) for More ❯
Boston, Massachusetts, United States Hybrid/Remote Options
Verndale
learning. What We're Looking For 2-4+ years of experience in building production-grade data pipelines and data systems. Strong proficiency in SQL, Python, or Java/Scala, with experience in distributed data frameworks (Apache Spark, Presto, EMR) Experience with modern data platforms (Snowflake, BigQuery, Redshift, Databricks, Synapse) Hands-on skills with streaming technologies (Kafka, Kinesis, Pub/ More ❯
and collaboratively within client teams. Desirable: Consulting experience or client-facing delivery background. Familiarity with tools such as dbt, Fivetran, Matillion , or similar. Programming skills in Python, Java, or Scala . Cloud certifications (SnowPro, Databricks Certified, AWS/Azure/GCP). Knowledge of DataOps, CI/CD , and infrastructure-as-code concepts. What’s on Offer Hybrid working model More ❯
City of London, London, United Kingdom Hybrid/Remote Options
Az-Tec Talent
and collaboratively within client teams. Desirable: Consulting experience or client-facing delivery background. Familiarity with tools such as dbt, Fivetran, Matillion , or similar. Programming skills in Python, Java, or Scala . Cloud certifications (SnowPro, Databricks Certified, AWS/Azure/GCP). Knowledge of DataOps, CI/CD , and infrastructure-as-code concepts. What’s on Offer Hybrid working model More ❯
with demonstrated experience writing complex queries, optimizing performance, and working across large-scale relational and cloud data environments. • Strong programming skills in at least one modern language (e.g., Python, Scala, or Java). • Experience with both relational and non-relational databases (e.g., PostgreSQL, SQL Server, DynamoDB). • Experience with cloud data platforms such as AWS (S3, Redshift, Glue), Azure, Snowflake More ❯
in big data technologies At least 1 year experience with cloud computing (AWS, Microsoft Azure, Google Cloud) Preferred Qualifications: 7+ years of experience in application development including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka More ❯
in big data technologies At least 1 year experience with cloud computing (AWS, Microsoft Azure, Google Cloud) Preferred Qualifications: 7+ years of experience in application development including Python, SQL, Scala, or Java 4+ years of experience with a public cloud (AWS, Microsoft Azure, Google Cloud) 4+ years experience with Distributed data/computing tools (Spark, MapReduce, Hadoop, Hive, EMR, Kafka More ❯
deployment in secure and scalable environments to include AI/ML frameworks such as TensorFlow, PyTorch, or scikit-learn. Proven expertise in programming languages such as Python, Java, or Scala, with demonstrated experience in software engineering practices (e.g., version control, CI/CD pipelines, containerization). Experience building and optimizing data pipelines, ETL processes, and real-time streaming solutions using More ❯
on proficiency with modern data technologies such as Spark, Kafka, Airflow, or dbt. Strong SQL skills and experience with cloud platforms (Azure preferred). Solid programming background in Python, Scala, or Java. Knowledge of data warehousing solutions (e.g. Snowflake, BigQuery, Redshift). Strong understanding of data governance, security, and compliance (experience within financial services is a plus). Leadership experience More ❯
London, South East, England, United Kingdom Hybrid/Remote Options
CV TECHNICAL LTD
on proficiency with modern data technologies such as Spark, Kafka, Airflow, or dbt. Strong SQL skills and experience with cloud platforms (Azure preferred). Solid programming background in Python, Scala, or Java. Knowledge of data warehousing solutions (e.g. Snowflake, BigQuery, Redshift). Strong understanding of data governance, security, and compliance (experience within financial services is a plus). Leadership experience More ❯
in Computer Science, Engineering, or a related field. Proven experience as a Data Engineer, with a focus on big data technologies. Strong proficiency in programming languages such as Python, Scala, or Java. Extensive experience with data warehousing, ETL processes, and data modeling. Experience with major cloud providers (e.g., AWS, GCP, Azure) and their data storage and processing services. Hands-on More ❯
of various Data repositories, Databases, ETL and BI tools and data assets hosted on Hadoop or cloud platforms Agile development experience & collaborating with data engineering teams Familiarity with Spark, Scala and distributed data processing platforms will be a big advantage Ability to work with tools like Jupyter/Databricks notebooks to interactively work with data sets Exusia, Inc. is committed More ❯
Machine Learning fundamentals and strong knowledge in a specific domain (e.g., Computer Vision, Deep Learning, NLP). Expert coding skills in Python and at least one of these languages: Scala, C++, Java. Hands-on experience with database management languages (e.g., SQL, PostgreSQL). Hands-on experience in cloud-based infrastructures (AWS/GCP/Azure). Hands-on experience with More ❯
models. Good understanding of ML fundamentals and strong knowledge in a domain (e.g., Computer Vision, Deep Learning, NLP). Expert coding skills in Python and at least one of Scala, C++, Java. Hands on experience with SQL, PostgreSQL and other database languages. Experience with cloud based infrastructures (AWS/GCP/Azure). Knowledge of Unix command line and DevOps More ❯