influence and establish best practices in data engineering processes. Self-motivated with a growth mindset and an eagerness to stay updated on emerging technologies. Apache Kafka Expertise : Proven experience designing, developing, and managing Kafka-based data pipelines . Good understanding of Kafka Streams , Connect, or the Confluent Kafka platform. … optimizing Snowflake schemas for efficient querying. Implementing ETL/ELT pipelines to load and transform data in Snowflake. Big Data Processing Frameworks : Familiarity with ApacheSpark , Hadoop, or other distributed data processing frameworks. Data Governance and Compliance : Understanding of data governance principles , security policies, and compliance standards (e.g. More ❯
City of London, London, United Kingdom Hybrid / WFH Options
McCabe & Barton
ideal candidate with have expertise in some of the following: Python, SQL, Scala, and Java for data engineering. Strong experience with big data tools (ApacheSpark, Hadoop, Databricks, Dask) and cloud platforms (AWS, Azure, GCP). Proficient in data modelling (relational, NoSQL, dimensional) and DevOps automation (Docker, Kubernetes More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
continuous improvement One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
experience as a Data Engineer with a strong background in data pipelines. Proficiency in Python, Java, or Scala, and big data technologies (e.g., Hadoop, Spark, Kafka). Experience with Databricks, Azure AI Services, and cloud platforms (AWS, Google Cloud, Azure). Solid understanding of SQL and NoSQL databases. Strong More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
deliverables. Strong background in data lake and data warehouse design, including data modeling and partitioning strategies Advanced proficiency in ETL tools (e.g., Talend, Informatica, Apache Airflow) and orchestration frameworks Extensive experience with cloud data ecosystems (AWS, Azure, GCP) and containerization (e.g., Docker, Kubernetes) In-depth knowledge of CI/… organizational goals One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
with SQL and Databricks for data transformations, analytics, and real-time processing Strong background in ETL processes and data integration tools (e.g., Talend, Informatica, Apache Airflow) Proficiency in Python scripting/programming language and SQL optimization Strong background in Agile project delivery, including sprint planning, task tracking, and collaboration … across departments One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
Preferred Qualifications: DOD 8570 IAT Level II Certification may be required (GSEC, GICSP, CND, CySA+, Security+ CE, SSCP or CCNA-Security). Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
Oakland, California, United States Hybrid / WFH Options
PG&E Corporation
in SQL and database management. Experience with cloud platforms such as AWS, Azure, or Google Cloud. Familiarity with Big Data technologies such as Hadoop, Spark, and Kafka is a plus. Excellent problem-solving and analytical skills. Strong communication and collaboration abilities. Experience with Python or other scripting languages. Knowledge More ❯
ensure high availability and accessibility. Experience & Skills : Strong experience in data engineering. At least some commercial hands-on experience with Azure data services (e.g., ApacheSpark, Azure Data Factory, Synapse Analytics). Proven experience in leading and managing a team of data engineers. Proficiency in programming languages such More ❯
london, south east england, united kingdom Hybrid / WFH Options
DATAHEAD
ensure high availability and accessibility. Experience & Skills : Strong experience in data engineering. At least some commercial hands-on experience with Azure data services (e.g., ApacheSpark, Azure Data Factory, Synapse Analytics). Proven experience in leading and managing a team of data engineers. Proficiency in programming languages such More ❯
Snowflake. Understanding of cloud platform infrastructure and its impact on data architecture. Data Technology Skills: A solid understanding of big data technologies such as ApacheSpark, and knowledge of Hadoop ecosystems. Knowledge of programming languages such as Python, R, or Java is beneficial. Exposure to ETL/ELT More ❯
Coalville, Leicestershire, East Midlands, United Kingdom Hybrid / WFH Options
Ibstock PLC
and BI solutions. Ensure data accuracy, integrity, and consistency across the data platform. Knowledge, Skills and Experience: Essentia l Strong expertise in Databricks and ApacheSpark for data engineering and analytics. Proficient in SQL and Python/PySpark for data transformation and analysis. Experience in data lakehouse development More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
technical audiences. One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
East London, London, United Kingdom Hybrid / WFH Options
Asset Resourcing
programming languages such as Python or Java. Understanding of data warehousing concepts and data modeling techniques. Experience working with big data technologies (e.g., Hadoop, Spark) is an advantage. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Responsibilities: Design, build and maintain efficient and scalable data pipelines More ❯
programming languages such as Python or Java. Understanding of data warehousing concepts and data modeling techniques. Experience working with big data technologies (e.g., Hadoop, Spark) is an advantage. Excellent problem-solving and analytical skills. Strong communication and collaboration skills. Benefits Enhanced leave - 38 days inclusive of 8 UK Public More ❯
unstructured datasets. Engineering best practices and standards. Experience with data warehouse software (e.g. Snowflake, Google BigQuery, Amazon Redshift). Experience with data tools: Hadoop, Spark, Kafka, etc. Code versioning (Github integration and automation). Experience with scripting languages such as Python or R. Working knowledge of message queuing and … stream processing. Experience with ApacheSpark or Similar Technologies. Experience with Agile and Scrum Technologies. Familiarity with dbt and Airflow is an advantage. Experience working in a start-up or scale up environment. Experience working in the fields of financial technology, traditional financial services, or blockchain/cryptocurrency. More ❯
quality of the work you deliver Furthermore, you have experience in: working with AWS developing applications in a Kubernetes environment developing batch jobs in ApacheSpark (pyspark or Scala) and scheduling them in an Airflow environment developing streaming applications for Apache Kafka in Python or Scala working More ❯
Newcastle Upon Tyne, Tyne and Wear, North East, United Kingdom Hybrid / WFH Options
Client Server
Data Engineer (Python Spark SQL) *Newcastle Onsite* to £70k Do you have a first class education combined with Data Engineering skills? You could be progressing your career at a start-up Investment Management firm that have secure backing, an established Hedge Fund client as a partner and massive growth … scientific discipline, backed by minimum A A B grades at A-level You have commercial Data Engineering experience working with technologies such as SQL, ApacheSpark and Python including PySpark and Pandas You have a good understanding of modern data engineering best practices Ideally you will also have … will earn a competitive salary (to £70k) plus significant bonus and benefits package. Apply now to find out more about this Data Engineer (Python Spark SQL) opportunity. At Client Server we believe in a diverse workplace that allows people to play to their strengths and continually learn. We're More ❯
services experience is desired but not essential. API development (FastAPI, Flask) Tech stack : Azure, Python, Databricks, Azure DevOps, ChatGPT, Groq, Cursor AI, JavaScript, SQL, ApacheSpark, Kafka, Airflow, Azure ML, Docker, Kubernetes and many more. Role Overview: We are looking for someone who is as comfortable developing AI More ❯
london, south east england, united kingdom Hybrid / WFH Options
Aventis Solutions
services experience is desired but not essential. API development (FastAPI, Flask) Tech stack : Azure, Python, Databricks, Azure DevOps, ChatGPT, Groq, Cursor AI, JavaScript, SQL, ApacheSpark, Kafka, Airflow, Azure ML, Docker, Kubernetes and many more. Role Overview: We are looking for someone who is as comfortable developing AI More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
team collaboration One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
functional teams. One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with ApacheSpark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra More ❯
Arlington, Virginia, United States Hybrid / WFH Options
Amazon
equivalent experience - 3+ years of experience with data warehouse architecture, ETL/ELT tools, data engineering, and large-scale data manipulation using technologies like Spark, EMR, Hive, Kafka, and Redshift - Experience with relational databases, SQL, and performance tuning, as well as software engineering best practices for the development lifecycle More ❯
Fairfax, Virginia, United States Hybrid / WFH Options
CGI
SAS, SPSS, and Oracle); build analytical solutions using programming languages (e.g., R, Python, SAS), and programming libraries (e.g., Python SciKit, R Caret, PostgreSQL MADlib, ApacheSpark MLlib). Ensure data integrity, consistency, and quality throughout the ETL process. Collect and transform structured, unstructured, relational, and NoSQL data using … insights create business impact. Experience with cloud-based ETL solutions (e.g., AWS Glue, Azure Data Factory). Knowledge of Big Data technologies (e.g., Hadoop, Spark). Certification in relevant ETL or data integration technologies. Bachelors Degree in data science, mathematics, statistics, economics, computer science, engineering, or other related business … Learning opportunities and tuition assistance Wellness and Well-being programs Due to the nature of this government contract, US Citizenship is required. Skills: Adobe Spark Hadoop Ecosystem (HDFS) What you can expect from us: Together, as owners, lets turn meaningful insights into action. Life at CGI is rooted in More ❯