City of London, London, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
Be Doing You'll be a key contributor to the development of a next-generation data platform, with responsibilities including: Designing and implementing scalable data pipelines using Python and Apache Spark Building and orchestrating workflows using AWS services such as Glue , Lambda , S3 , and EMR Serverless Applying best practices in software engineering: CI/CD , version control , automated testing … and modular design Supporting the development of a lakehouse architecture using ApacheIceberg Collaborating with product and business teams to deliver data-driven solutions Embedding observability and quality checks into data workflows Participating in code reviews, pair programming, and architectural discussions Gaining domain knowledge in financial data and sharing insights with the team What They're Looking For … for experience with type hints, linters, and testing frameworks like pytest) Solid understanding of data engineering fundamentals: ETL/ELT, schema evolution, batch processing Experience or strong interest in Apache Spark for distributed data processing Familiarity with AWS data tools (e.g., S3, Glue, Lambda, EMR) Strong communication skills and a collaborative mindset Comfortable working in Agile environments and engaging More ❯
London, South East, England, United Kingdom Hybrid / WFH Options
Tenth Revolution Group
Be Doing You'll be a key contributor to the development of a next-generation data platform, with responsibilities including: Designing and implementing scalable data pipelines using Python and Apache Spark Building and orchestrating workflows using AWS services such as Glue , Lambda , S3 , and EMR Serverless Applying best practices in software engineering: CI/CD , version control , automated testing … and modular design Supporting the development of a lakehouse architecture using ApacheIceberg Collaborating with product and business teams to deliver data-driven solutions Embedding observability and quality checks into data workflows Participating in code reviews, pair programming, and architectural discussions Gaining domain knowledge in financial data and sharing insights with the team What They're Looking For … for experience with type hints, linters, and testing frameworks like pytest) Solid understanding of data engineering fundamentals: ETL/ELT, schema evolution, batch processing Experience or strong interest in Apache Spark for distributed data processing Familiarity with AWS data tools (e.g., S3, Glue, Lambda, EMR) Strong communication skills and a collaborative mindset Comfortable working in Agile environments and engaging More ❯
Arlington, Virginia, United States Hybrid / WFH Options
STR
the following software/tools: Big Data tools: e.g. Hadoop, Spark, Kafka, ElasticSearch AWS: Athena, RDB, AWS credentials from Cloud Practitioner to Solutions Architect Data Lakes: e.g. Delta Lake, Apache Hudi, ApacheIceberg Distributed SQL interfaces: e.g. Apache Hive, Presto/Trino, Spark Data pipeline and workflow management tools: e.g Luigi, Airflow Dashboard frontends: e.g. Grafana More ❯
Chantilly, Virginia, United States Hybrid / WFH Options
The DarkStar Group
rather huge and includes Python (Pandas, numpy, scipy, scikit-learn, standard libraries, etc.), Python packages that wrap Machine Learning (packages for NLP, Object Detection, etc.), Linux, AWS/C2S, Apache NiFi, Spark, pySpark, Hadoop, Kafka, ElasticSearch, Solr, Kibana, neo4J, MariaDB, Postgres, Docker, Puppet, and many others. Work on this program takes place in Chantilly, VA, McLean, VA and in … standards. Develop and deliver documentation for each project including ETL mappings, code use guide, code location and access instructions. Design and optimize Data Pipelines using tools such as Spark, ApacheIceberg, Trino, OpenSearch, EMR cloud services, NiFi and Kubernetes containers Ensure the pedigree and provenance of the data is maintained such that the access to data is protected More ❯
Herndon, Virginia, United States Hybrid / WFH Options
The DarkStar Group
rather huge and includes Python (Pandas, numpy, scipy, scikit-learn, standard libraries, etc.), Python packages that wrap Machine Learning (packages for NLP, Object Detection, etc.), Linux, AWS/C2S, Apache NiFi, Spark, pySpark, Hadoop, Kafka, ElasticSearch, Solr, Kibana, neo4J, MariaDB, Postgres, Docker, Puppet, and many others. Work on this program takes place in Chantilly, VA, McLean, VA and in … standards. Develop and deliver documentation for each project including ETL mappings, code use guide, code location and access instructions. Design and optimize Data Pipelines using tools such as Spark, ApacheIceberg, Trino, OpenSearch, EMR cloud services, NiFi and Kubernetes containers Ensure the pedigree and provenance of the data is maintained such that the access to data is protected More ❯
MySQL Exposure to Docker, Kubernetes, AWS, Helm, Terraform, Vault, Grafana, ELK Stack, New Relic Relevant experience in the maintenance of data APIs and data lake architectures, including experience with ApacheIceberg, Trino/Presto, Clickhouse, Snowflake, BigQuery. Master's degree in Computer Science or Engineering-related field Equal Opportunity Employer As an Equal Opportunity Employer, qualified applicants will More ❯
Falls Church, Virginia, United States Hybrid / WFH Options
Rackner
years of software engineering experience (backend, API, or full-stack) - Python, Java, or C# expertise - Experience with REST APIs (FastAPI, AWS Lambda), OpenAPI, and data pipelines (dbt, Airflow, Spark, Iceberg) - Knowledge of FHIR, OMOP, HL7, CDA, and federal compliance frameworks Bonus Experience: - DHA, VA, or federal healthcare IT programs - OCR/NLP/AI-ML workflows - AWS GovCloud (IL5 More ❯
running such systems in production. Strong coding skills in Java/Python and familiarity with CI/CD. Hands-on with some of: Kafka/Flink, Spark, Delta/Iceberg, Kubernetes, NoSQL/columnar stores. Proven ability to work independently, make sound tradeoffs, and deliver quality outcomes with minimal supervision. Solid debugging, performance analysis, and system design skills. Nice More ❯
or equivalent) and in infra-as-code, CI/CD, and containerized environments. Hands-on deep internal expertise in several of the following: Kafka/Flink, Spark,, Delta/Iceberg, GraphQL/REST APIs, RDBMS/NoSQL, Kubernetes, Airflow. Experience building both streaming and batch data platforms, improving reliability, quality, and developer velocity. Demonstrated ability to mentor senior engineers More ❯
Manchester, Lancashire, England, United Kingdom Hybrid / WFH Options
Lorien
passionate about building scalable, cloud-native data platforms. You'll be a key player in a growing team, helping to shape the future of data infrastructure using AWS, PySpark, Iceberg, and more. From designing high-performance pipelines to supporting a full-scale migration from SQL Server to AWS, this role offers the chance to work on real-time data More ❯