learning libraries in one or more programming languages. Keen interest in some of the following areas: Big Data Analytics (e.g. Google BigQuery/BigTable, Apache Spark), Parallel Computing (e.g. Apache Spark, Kubernetes, Databricks), Cloud Engineering (AWS, GCP, Azure), Spatial Query Optimisation, Data Storytelling with (Jupyter) Notebooks, Graph Computing More ❯
Washington, Washington DC, United States Hybrid / WFH Options
SMX
SMX is seeking a talented Data Engineer (Python) with expertise in ETL (Extract, Transform, Load) processes and Apache Airflow. The candidate will be responsible for designing and implementing robust and efficient data pipelines, ensuring high data quality, and contributing to the continuous improvement of our data management practices. This … is a remote position supporting a Washington, DC based team. Essential Duties & Responsibilities: Design, develop, and maintain ETL processes using Python and Apache Airflow. Collaborate with data analysts and other stakeholders to understand and meet their data requirements. Develop and implement data validation processes to ensure high data quality. … Skills & Experience: Proficiency in Python: Strong understanding of Python programming language. Experience with Python libraries and frameworks like Pandas, NumPy, and Django. Expertise in Apache Airflow: Experience in designing, building, and maintaining data pipelines using Apache Airflow. Knowledge of Airflow's architecture, including DAGs and Operators. ETL Processes More ❯
chelmsford, east anglia, united kingdom Hybrid / WFH Options
EMBL-EBI
services, including microservice deployment Developing and maintaining data pipelines to process and manage large-scale structural biology data Working with workflow orchestration tools like Apache Airflow and Nextflow To liaise with the EMBL-EBI Technical Services Cluster To support the development and deployment of other scientific software The successful … Proficiency in containerisation Proficiency in relational databases (Oracle, PostgreSQL) Experience in developing and maintaining data pipelines Hands-on experience with workflow orchestration tools (e.g., Apache Airflow, Nextflow) Strong interpersonal and communication skills Proficiency in oral and written English You may also have PhD in computer science, IT or a … related field, or in bioinformatics with a demonstrated IT expertise Experience in using Kubernetes Experience with web servers (Apache/Nginx) Hands-on experience with CI/CD (GitLab CI/GitHub Actions) Familiarity with networking Familiarity with Java Knowledge of, or affinity with, structural biology and bioinformatics Experience More ❯
saffron walden, east anglia, united kingdom Hybrid / WFH Options
EMBL-EBI
services, including microservice deployment Developing and maintaining data pipelines to process and manage large-scale structural biology data Working with workflow orchestration tools like Apache Airflow and Nextflow To liaise with the EMBL-EBI Technical Services Cluster To support the development and deployment of other scientific software The successful … Proficiency in containerisation Proficiency in relational databases (Oracle, PostgreSQL) Experience in developing and maintaining data pipelines Hands-on experience with workflow orchestration tools (e.g., Apache Airflow, Nextflow) Strong interpersonal and communication skills Proficiency in oral and written English You may also have PhD in computer science, IT or a … related field, or in bioinformatics with a demonstrated IT expertise Experience in using Kubernetes Experience with web servers (Apache/Nginx) Hands-on experience with CI/CD (GitLab CI/GitHub Actions) Familiarity with networking Familiarity with Java Knowledge of, or affinity with, structural biology and bioinformatics Experience More ❯
Saffron Walden, Essex, South East, United Kingdom Hybrid / WFH Options
EMBL-EBI
services, including microservice deployment Developing and maintaining data pipelines to process and manage large-scale structural biology data Working with workflow orchestration tools like Apache Airflow and Nextflow To liaise with the EMBL-EBI Technical Services Cluster To support the development and deployment of other scientific software The successful … Proficiency in containerisation Proficiency in relational databases (Oracle, PostgreSQL) Experience in developing and maintaining data pipelines Hands-on experience with workflow orchestration tools (e.g., Apache Airflow, Nextflow) Strong interpersonal and communication skills Proficiency in oral and written English You may also have PhD in computer science, IT or a … related field, or in bioinformatics with a demonstrated IT expertise Experience in using Kubernetes Experience with web servers (Apache/Nginx) Hands-on experience with CI/CD (GitLab CI/GitHub Actions) Familiarity with networking Familiarity with Java Knowledge of, or affinity with, structural biology and bioinformatics Experience More ❯
have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark. You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative … Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark is a plus Where You'll Be This role is based in London (UK). We offer you the flexibility to More ❯
have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark. You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative … Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark is a plus Where You'll Be This role is based in London (UK). We offer you the flexibility to More ❯
have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative … Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark is a plus Where You'll Be This role is based in London (UK) We offer you the flexibility to work More ❯
companies where years-long behemoth projects are the norm, our projects are fast-paced, typically 2 to 4 months long. Most are delivered using Apache Spark/Databricks on AWS/Azure and require you to directly manage the customer relationship alone or in collaboration with a Project Manager. … at DATAPAO, meaning that you'll get access to Databricks' public and internal courses to learn all the tricks of Distributed Data Processing, MLOps, Apache Spark, Databricks, and Cloud Migration from the best. Additionally, we'll pay for various data & cloud certifications, you'll get dedicated time for learning … seniority level during the selection process. About DATAPAO At DATAPAO, we are delivery partners and the preferred training provider for Databricks, the creators of Apache Spark. Additionally, we are Microsoft Gold Partners in delivering cloud migration and data architecture on Azure. Our delivery partnerships enable us to work in More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Epsilon Inc
with SQL and Databricks for data transformations, analytics, and real-time processing Strong background in ETL processes and data integration tools (e.g., Talend, Informatica, Apache Airflow) Proficiency in Python scripting/programming language and SQL optimization Strong background in Agile project delivery, including sprint planning, task tracking, and collaboration … across departments One or more of the following certifications are desired: AWS Certified Developer, Databricks, Agile/Scrum, Python Programmer Preferred Qualifications: Familiarity with Apache Spark or comparable distributed data processing frameworks, preferably for large-scale data transformations and analytics. Working knowledge of data governance platforms (e.g., Collibra) and More ❯
Alexandria, Virginia, United States Hybrid / WFH Options
CGI
Vanity URLs, Asset, Dynamic Media and Tag management. AEM hands-on programming in Web Technologies like JSP, JavaScript, XML, XPATH, AJAX, jQuery, Adobe APIs, Apache Sling, OSGi bundles, Apache Web Server. Adobe AEM, Commerce, Magento 2 Certification or equivalent experience is highly preferred. Strong understanding of business processes More ❯
Tampa, Florida, United States Hybrid / WFH Options
LTIMindtree
in Python Programming Python Institute Certified Professional in Python Programming 1 Python Institute Certified Professional in Python Programming 2 Databricks Certified Associated Developer for Apache Spark Skills Mandatory Skills : Apache Spark, Big Data Hadoop Ecosystem, Data Architecture, Python More ❯
quality of the work you deliver Furthermore, you have experience in: working with AWS developing applications in a Kubernetes environment developing batch jobs in Apache Spark (pyspark or Scala) and scheduling them in an Airflow environment developing streaming applications for Apache Kafka in Python or Scala working with More ❯
Fairfax, Virginia, United States Hybrid / WFH Options
CGI
deploying, and managing interactive dashboards and reports using Power BI. Understanding of Snowflake's architecture, including virtual warehouses, data sharing, and clustering. Proficiency in Apache Spark and Databricks for big data processing and analytics. Familiarity with Delta Lake for managing large-scale data in a lake-house architecture. Strong … tuition assistance Wellness and Well-being programs Due to the nature of this government contract, US Citizenship is required. Skills: Oracle PowerShell PowerShell Snowflake Apache Spark What you can expect from us: Together, as owners, lets turn meaningful insights into action. Life at CGI is rooted in ownership, teamwork More ❯
Lexington, Massachusetts, United States Hybrid / WFH Options
Equiliem
Computer Science. Recent graduates or candidates without a Bachelor's degree considered with clear evidence of significant outside-of-classroom experience. • Experience with the Apache Maven or Gradle build system. • Ability to understand front-end source code written in React or similar frameworks. Provide guidance to less experienced front … and environments, such as Pandas, TensorFlow, and Jupyter Notebook. • Broad knowledge of the general features, capabilities, and trade-offs of common data warehouse (e.g. Apache Hadoop); workflow orchestration (e.g. Apache Beam); data extract, transform and load (ETL); and stream processing (e.g. Kafka) technologies. Hands-on experience with several More ❯
sync/async methods. Expertise in databases, database architecture, SQL, and stored procedures. Strong background in data integration technologies such as Azure Data Factory, Apache Airflow, and Databricks AutoLoader. Proficiency in building CI/CD pipelines and implementing infrastructure as code. Experience with event-driven data flows and message … aggregation technologies (Event Hub, Apache Kafka). A solid understanding of current and emerging technologies and how they can deliver business value. Desirable experience with data matching processes. Strong ability to translate business strategies into technical solutions and business requirements into technical designs. If you're passionate about driving More ❯
Leeds, Yorkshire, United Kingdom Hybrid / WFH Options
Oliver James Associates Ltd
sync/async methods. Expertise in databases, database architecture, SQL, and stored procedures. Strong background in data integration technologies such as Azure Data Factory, Apache Airflow, and Databricks AutoLoader. Proficiency in building CI/CD pipelines and implementing infrastructure as code. Experience with event-driven data flows and message … aggregation technologies (Event Hub, Apache Kafka). A solid understanding of current and emerging technologies and how they can deliver business value. Desirable experience with data matching processes. Strong ability to translate business strategies into technical solutions and business requirements into technical designs. If you're passionate about driving More ❯
Boston, Massachusetts, United States Hybrid / WFH Options
Digital Prospectors
workflows and propose streamlined infrastructure solutions. • Troubleshoot Linux-based desktop and server systems. • Provide technical support for infrastructure services such as NFS, DHCP, DNS, Apache, MySQL, Squid, and mediaWiki. • Contribute to AWS GovCloud infrastructure, including enhancements and integration using IaC tools such as Terraform, Kubernetes, or Packer. • Participate in … or Packer). • Experience integrating Linux in environments where Microsoft Windows is the dominant client OS. • Ability to troubleshoot and support services such as Apache, MySQL, NFS, DNS, DHCP, mediaWiki, and Squid. • Hands-on experience with automated configuration tools, preferably Ansible. • Experience with Debian/Ubuntu distributions on both More ❯
Edinburgh, Scotland, United Kingdom Hybrid / WFH Options
TalentHawk
early and providing strategic guidance. Support the ongoing development of integration strategies involving Managed File Transfer solutions (e.g., GoAnywhere) and data orchestration platforms (e.g., Apache Airflow). Provide hands-on support and detailed guidance on particularly complex integration designs where necessary. Maintain current knowledge of industry trends, technology advancements … and migrations. Familiarity with IBM Maximo asset management platform. Knowledge and experience with Managed File Transfer solutions (e.g., GoAnywhere). Understanding and experience with Apache Airflow orchestration platform. Strong grasp of integration best practices, security considerations, and data flow management. Ability to work collaboratively across distributed teams and multiple More ❯
data governance and best practice. ·Become an SME on the design, development, and deployment of data ETL pipelines (using Azure Data Factory, Azure Synapse, Apache Spark and other technologies) to access, combine and transform data from on-prem and cloud-based sources. ·Ensure that all data pipelines are developing … and balance the need for delivery over scalability Experience & Skills Required ·Proven track record of developing data pipelines and products using Azure, Azure Synapse, Apache Spark, DevOps, Snowflake, Databricks and Fabric. ·High level of coding proficiency in SQL and Python. ·A good level of experience of Data Modelling and More ❯
Leeds, Yorkshire, United Kingdom Hybrid / WFH Options
Sportserve
Sportserve forms part of a remarkable group of B2C sports betting and B2B sportsbook technology companies, focused on delivering first class sports betting experiences and casino products for our users worldwide. Along with Sportion, TechSpirit, Standard Focus and Sportelligent, we More ❯
City of London, London, United Kingdom Hybrid / WFH Options
I3 Resourcing Limited
Data Platform Engineer - SSIS & T-SQL, Data Factory - Hybrid Data Platform Engineer SSIS & T-SQL, Data Factory, Databricks/Apache Spark London Insurance Market City, London/Hybrid (3 days per week in the office) Permanent £85,000 per annum + benefits + bonus PLEASE ONLY APPLY IF YOU … data function in a London Market Insurance setting Sound understanding of data warehousing concepts ETL/ELTs - SSIS & T-SQL, Data Factory, Databricks/Apache Spark Data modelling Strong communication skills and able to build relationships and trust with stakeholders Data Platform Engineer SSIS & T-SQL, Data Factory, Databricks …/Apache Spark London Insurance Market City, London/Hybrid (3 days per week in the office) Permanent £85,000 per annum + benefits + bonus More ❯
Bristol, Gloucestershire, United Kingdom Hybrid / WFH Options
Apacheix
disciplinary teams that may be from multiple organisations Working with your clients to help them deliver their goals successfully, whilst demonstrating and upholding the Apache values Continuously seeking to develop your cyber security skills, learning from a range of sources, including formal training, learning from experience, and coaching and … mentoring Coaching and mentoring more junior colleagues to help transfer your knowledge and skills, either with Apache or with our clients Building and maintaining excellent client relationships Contributing to business development activities and the winning of new client contracts Experience you'll need Experience in delivering Defence Cyber, Enterprise … working Work from home or our Bristol offices The latest secure tech Investment in personal development Vibrant social scene Relocation bonus for Graduates Why Apache iX? Our growing team brings a wealth of experience from across the defence and security sector, and we pride ourselves in delivering the highest More ❯
driving business value through ML Company first focus and collaborative individuals - we work better when we work together. Preferred Experience working with Databricks and Apache Spark Preferred Experience working in a customer-facing role About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide … data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our More ❯
driving business value through ML Company first focus and collaborative individuals - we work better when we work together. Preferred Experience working with Databricks and Apache Spark Preferred Experience working in a customer-facing role About Databricks Databricks is the data and AI company. More than 10,000 organizations worldwide … data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn, and Facebook. Benefits At Databricks, we strive to provide comprehensive benefits More ❯