basics like HTTP, HTML, REST, CSS, and JavaScript Preferred Technical Skills: Proficient in at least one scripting language like Python or Perl Familiar with web server configuration on NGINX, Apache, or similar Familiar with a JavaScript framework like Backbone, Angular, Ember or Vue.js Familiar with Bootstrap and Less Experience with NoSQL storage like ElasticSearch, Accumulo, MongoDB, HBase is a More ❯
as required; comfortable building multi page Web Applications from scratch. Expertise with Application Server integration; JBoss 7, SpringBoot or later preferred. Proficient in developing microservices with SpringBoot Knowledge of Apache Web Server preferred. Database Skills with working knowledge of Structured Query Language (e.g. SQL/NoSQL commands and queries). 2+ years Working with Oracle, MySQL, MS SQL and … as required; comfortable building multi page Web Applications from scratch. Expertise with Application Server integration; JBoss 7, SpringBoot or later preferred. Proficient in developing microservices with SpringBoot Knowledge of Apache Web Server preferred. Database Skills with working knowledge of Structured Query Language (e.g. SQL/NoSQL commands and queries). 2+ years Working with Oracle, MySQL, MS SQL and More ❯
Bethesda, Maryland, United States Hybrid / WFH Options
Sunayu, LLC
us even more if you have some of these skills: Active Security+ certification Active CE certification Ability to obtain Privileged User Account (PUA) certification by the Government GraphQL Service Apache NiFi Experience with systems at scale Kafka or other message passing systems Experience of a production/enterprise system Appian low-code application development platform Experience with any of More ❯
Bethesda, Maryland, United States Hybrid / WFH Options
Leidos
us even more if you have some of these skills: Active Security+ certification Active CE certification Ability to obtain Privileged User Account (PUA) certification by the Government GraphQL Service Apache NiFi Experience with systems at scale Kafka or other message passing systems Experience of a production/enterprise system Appian low-code application development platform Experience with any of More ❯
technical direction to a growing team of developers globally. The platform is a Greenfield build using standard modern technologies such as Java, Spring Boot, Kubernetes, Kafka, MongoDB, RabbitMQ, Solace, Apache Ignite. The platform runs in a hybrid mode both on-premise and in AWS utilising technologies such as EKS, S3, FSX. The main purpose of this role is to More ❯
Responsibilities: Develop, optimize, and maintain data ingest flows using Apache Kafka, Apache Nifi and MySQL/PostGreSQL Develop within the components in the AWS cloud platform using services such as RedShift, SageMaker, API Gateway, QuickSight, and Athena Communicate with data owners to set up and ensure configuration parameters Document SOP related to streaming configuration, batch configuration or API … machine learning techniques Strong understanding of programming languages like Python, R, and Java Expertise in building modern data pipelines and ETL (extract, transform, load) processes using tools such as Apache Kafka and Apache Nifi Proficient in programming languages like Java, Scala, or Python Experience or expertise using, managing, and/or testing API Gateway tools and Rest APIs More ❯
as Java, C, C++ for distributed systems, with proficiency in networking, multi-threading and implementation of REST APIs Experience with the Spring framework, messaging frameworks (Kafka, RabbitMQ), streaming analytics (Apache Flink, Spark), management of containerized applications (Kubernetes). Experience with Enabling tools (Git, Maven, Jira), DevOps (Bamboo, Jenkins, GitLab Cl/Pipelines), Continuous Monitoring (ELK Stack (ElasticSearch, Logstash and … Kibana), Nagios) Experience with Apache Hadoop, Apache Accumulo and Apache NiFi Well-grounded in Linux fundamentals and familiarity with scripting languages (e.g., Python, Ruby, Perl, BASH, etc.) Experience with AWS Cloud architecture, Infrastructure as Code (IaC), Cloud security, and Automation (AWS Lambda, CloudFormation)Benefits: Peraton offers enhanced benefits to employees working on this critical National Security program More ❯
In this role, you will be responsible for designing, building, and maintaining robust data pipelines and infrastructure on the Azure cloud platform. You will leverage your expertise in PySpark, Apache Spark, and Apache Airflow to process and orchestrate large-scale data workloads, ensuring data quality, efficiency, and scalability. If you have a passion for data engineering and a … data processing workloads Implement CI/CD pipelines for data workflows to ensure smooth and reliable deployments. Big Data & Analytics: Build and optimize large-scale data processing pipelines using Apache Spark and PySpark Implement data partitioning, caching, and performance tuning for Spark-based workloads. Work with diverse data formats (structured and unstructured) to support advanced analytics and machine learning More ❯
to-end, scalable data and AI solutions using the Databricks Lakehouse (Delta Lake, Unity Catalog, MLflow). Design and lead the development of modular, high-performance data pipelines using Apache Spark and PySpark. Champion the adoption of Lakehouse architecture (bronze/silver/gold layers) to ensure scalable, governed data platforms. Collaborate with stakeholders, analysts, and data scientists to … performance tuning, cost optimisation, and monitoring across data workloads. Mentor engineering teams and support architectural decisions as a recognised Databricks expert. Essential Skills & Experience: Demonstrable expertise with Databricks and Apache Spark in production environments. Proficiency in PySpark, SQL, and working within one or more cloud platforms (Azure, AWS, or GCP). In-depth understanding of Lakehouse concepts, medallion architecture More ❯
utilizing the Django web framework for the backends and React for developing the client facing portion of the application Create, extract, transform, and load (ETL) pipelines using Hadoop and Apache Airflow for various production big data sources to fulfill intelligence data availability requirements Automate retrieval of data from various sources via API and direct database queries for intelligence analysts … for military personnel Required Qualifications: Active TS/SCI Required 7-10 years experience Preferred Qualifications: Bachelor's degree in related field preferred Windows 7/10, MS Project Apache Airflow Python, Java, JavaScript, React, Flask, HTML, CSS, SQL, R, Docker, Kubernetes, HDFS, Postgres, Linux AutoCAD JIRA, Gitlab, Confluence About Us: IntelliBridge delivers IT strategy, cloud, cybersecurity, application, data More ❯
participating in sprint planning, daily stand-ups, and retrospectives. Performing code reviews, writing unit tests, and ensuring high-quality software is delivered. Supporting the integration of additional technologies, like Apache Spark, Kafka, and Kubernetes, as needed for system scalability and data handling. Troubleshooting and debugging production issues in a Linux environment. Staying updated on industry trends and best practices … start date) or be able to demonstrate AWS competence using the CLI and Boto Experience with the following Programming languages: C, C++, Java, JavaScript, Python Familiarity and knowledge of: Apache Spark, Kubernetes, Kafka, SIMP Project, Ansible, Docker, GIT, Linux, Suricata, Zeek, Kibana, Logstash, Elastic Search, Neo4J, PostgreSQL, Terraform, Artifactory, GitLab Experience or knowledge in the following areas: network engineering More ❯
Data Storage & Databases: SQL & NoSQL Databases: Experience with databases like PostgreSQL, MySQL, MongoDB, and Cassandra. Big Data Ecosystems: Hadoop, Spark, Hive, and HBase. Data Integration & ETL: Data Pipelining Tools: Apache NiFi, Apache Kafka, and Apache Flink. ETL Tools: AWS Glue, Azure Data Factory, Talend, and Apache Airflow. AI & Machine Learning: Frameworks: TensorFlow, PyTorch, Scikit-learn, Keras More ❯
and well-tested solutions to automate data ingestion, transformation, and orchestration across systems. Own data operations infrastructure: Manage and optimise key data infrastructure components within AWS, including Amazon Redshift, Apache Airflow for workflow orchestration and other analytical tools. You will be responsible for ensuring the performance, reliability, and scalability of these systems to meet the growing demands of data … pipelines , data warehouses , and leveraging AWS data services . Strong proficiency in DataOps methodologies and tools, including experience with CI/CD pipelines, containerized applications , and workflow orchestration using Apache Airflow . Familiar with ETL frameworks, and bonus experience with Big Data processing (Spark, Hive, Trino), and data streaming. Proven track record - You've made a demonstrable impact in More ❯
for performance, reliability, and scalability. Collaborate with engineers and analysts to understand data requirements and deliver high-quality data solutions. Implement and manage ETL processes using advanced tools like Apache Airflow, Spark, or similar. Ensure data quality and consistency through rigorous testing, validation, and governance practices. Deploy, monitor, and maintain data infrastructure in cloud environments (AWS, GCP, Azure). … Security+) Experience or expertise using, managing, and/or testing API Gateway tools and Rest APIs Experience or expertise configuring an LDAP client to connect to IPA Experience with Apache Hadoop and ETL Who we are: Reinventing Geospatial, Inc. (RGi) is a fast-paced small business that has the environment and culture of a start-up, with the stability More ❯
life-cycle management, and development tools: Hibernate, SpringBoot, ExtJS, AngularJS, Ansible, Swagger, Git, Subversion, Maven, Jenkins, Gradle, Nexus, Eclipse, IntelliJ, Ext-Js, JQuery, and D3. Cloud technologies: Pig, Hive, Apache Spark, Azure DataBricks, Storm, HBase, Hadoop Distributed File System, and MapReduce Open-source virtual machines and Cloud-based systems: OpenStack, OpenShift, Docker, Kubernetes, Microsoft Azure, and Amazon Web Services … Application servers: Node.js, Oracle Weblogic, Oracle Glassfish, JBoss and Apache Tomcat Data environments: Oracle, Sybase, PostGreSQL, MySQL, MongoDB, Splunk, and Elasticsearch/Logstash/Kibana (ELK), Neo4j Operating system environments: Windows, Linux, and MacOS Software requirements management: Atlassian Suite, EPICS, Story Board Data Science: Statistical Analysis, Data Visualization, Graph Analysis, Data Engineering This position is contingent on funding and More ❯
Columbia, South Carolina, United States Hybrid / WFH Options
Systemtec Inc
technologies and cloud-based technologies AWS Services, State Machines, CDK, Glue, TypeScript, CloudWatch, Lambda, CloudFormation, S3, Glacier Archival Storage, DataSync, Lake Formation, AppFlow, RDS PostgreSQL, Aurora, Athena, Amazon MSK, Apache Iceberg, Spark, Python ONSITE: Partially onsite 3 days per week (Tue, Wed, Thurs) and as needed. Standard work hours: 8:30 AM - 5:00 PM Required Qualifications of the More ❯
evolution Has experience (or strong interest) in building real-time or event-driven architectures ️ Modern Data Stack Includes: Python , SQL Snowflake , Postgres AWS (S3, ECS, Terraform) Airflow , dbt , Docker Apache Spark , Iceberg What they're looking for: Solid experience as a Senior/Lead/Principal Data Engineer, ideally with some line management or mentoring Proven ability to design More ❯
standards. Develop and deliver documentation for each project including ETL mappings, code use guide, code location and access instructions. Design and optimize Data Pipelines using tools such as Spark, Apache Iceberg, Trino, OpenSearch, EMR cloud services, NiFi and Kubernetes containers Ensure the pedigree and provenance of the data is maintained such that the access to data is protected Clean More ❯
Salary: 50.000 - 60.000 € per year Requirements: • 3+ years of hands-on experience as a Data Engineer working with Databricks and Apache Spark • Strong programming skills in Python, with experience in data manipulation libraries (e.g., PySpark, Spark SQL) • Experience with core components of the Databricks ecosystem: Databricks Workflows, Unity Catalog, and Delta Live Tables • Solid understanding of data warehousing principles More ❯
solutions using Databricks on Azure or AWS. Databricks Components : Proficient in Delta Lake, Unity Catalog, MLflow, and other core Databricks tools. Programming & Query Languages : Strong skills in SQL and Apache Spark (Scala or Python). Relational Databases : Experience with on-premises and cloud-based SQL databases. Data Engineering Techniques : Skilled in Data Governance, Architecture, Data Modelling, ETL/ELT More ❯
South East London, London, United Kingdom Hybrid / WFH Options
TEN10 SOLUTIONS LIMITED
and data validation techniques. Experience using test automation frameworks for data pipelines and ETL workflows Strong communication and stakeholder management skills. Nice-to-Have: Hands-on experience with Databricks , Apache Spark , and Azure Deequ . Familiarity with Big Data tools and distributed data processing. Experience with data observability and data quality monitoring. Proficiency with CI/CD tools like More ❯
models in close cooperation with our data science team Experiment in your domain to improve precision, recall, or cost savings Requirements Expert skills in Java or Python Experience with Apache Spark or PySpark Experience writing software for the cloud (AWS or GCP) Speaking and writing in English enables you to take part in day-to-day conversations in the More ❯
validation effort for a deliverable software solution Preferred Skills and Experience Digital Signal Processing Geolocation knowledge J2EE, Javascript, Bash Hibernate, SpringBoot, AngularJS, Ansible, Git, Subversion, Jenkins, Gradle Cloud technologies, Apache Spark, Azure, VM Docker, Container, Kubernetes Space Domain Awareness projects requiring software development CI/CD Pipeline Position is contingent upon successful contract award. Competitive salary based on experience More ❯
methodologies. Collaborating with stakeholders to define data strategies, implement data governance policies, and ensure data security and compliance. About you: Strong technical proficiency in data engineering technologies, such as Apache Airflow, ClickHouse, ETL tools, and SQL databases. Deep understanding of data modeling, ETL processes, data integration, and data warehousing concepts. Proficiency in programming languages commonly used in data engineering More ❯