real-time systems, and business management information systems) based upon documented requirements for the Data Transport System (DTS) • DTS products include but are not limited to: Cloud storage areas: ApacheAccumulo (Apache Zookeeper, ApacheHadoop), Oracle DMBS Real time streaming: Storm Distributed in-memory data Cache/Storage: Redis, Graph Compute engine/Query Interface apache Tinkerpop …/Gremlin. Rules Engine: JBoss Drools, Apache Camel, Spring Framework: used extensively to standardize/simplify configuration logic control, data access, security, Web-tier, etc. Candidates will: o Analyze user requirements to derive software design and performance requirements o Debug existing software and correct defects o Design and code new software or modify existing software to add new features … DoD) and other federal partners • The DTS portfolio encompasses transport streams, messages and files with content size ranging from bytes to Terabytes • Candidates should have experience writing analytics using Apache Hadoop, HDFS, and MapReduce • Experience processing large data sets or high-volume data ingest is a plus • Experience monitoring, maintaining and troubleshooting ApacheAccumulo, Apache Hadoop More ❯
languages, frameworks, and tools to support a broad range of applications Work with database technologies such as PostgreSQL, Redis, MySQL, and others Aid algorithm and data pipeline development in Apache Nifi Work Environment: Location: Colorado Springs, CO - 100% Onsite Travel Requirements: Minimal Working Hours: Standard Qualifications: Required: Security Clearance: Must have an active U.S. government Top Secret/SCI … which is something only a U.S. citizen can obtain Education: Bachelor's Degree in computer, information systems or related field Advanced proficiency in SQL and NoSQL databases Experience with ApacheAccumulo, Apache Hadoop Experience with Python Experience with Docker, AWS and/or Azure Hands-on experience with Apache Kafka, Apache NiFi Experience developing data … as xml, and protocol buffers (protobuf) Ability to work independently to research and solve customer pain points Desired: Master's Degree in Computer Science or related field Experience with Apache Artimis, ActiveMQ, or other IoT message brokers Working knowledge of MIL-STD-6016 and MIL-STD-3011 High level of curiosity and investigative mindset with an attention to detail More ❯
Software Engineer 3 to perform, amongst the Compute Team, within a massively parallel enterprise platform, built with Java on Free and Open-Source Software products including Kubernetes, Hadoop and Accumulo, to enable execution of data-intensive analytics on a managed infrastructure. The selected candidate will be a self-motivated Java developer who proactively completes tasks with a strong attention … and will be exposed to a variety of technologies depending on customer requirements. Required Skills: s Java programming for distributed systems, with experience in networking and multi-threading s Apache Hadoop s ApacheAccumulo s Apache NiFi s Agile development experience s Well-grounded in Linux fundamentals and knowledge in at least one scripting language (e.g. More ❯
Trip to either Nassau Bahamas, Singer Island Florida, Paradise Island Bahamas, or the Cambridge Hyatt Resort Desired Skills: • Proficient in Java • Comfortable working in a Linux environment • Experience with Apache Open Source Hadoop, Apache Open Source Accumulo, Apache Open Source NiFi • Familiarity with Context chaining and Graph theory • Experience with Containerization - Docker, Kubernetes • Experience with Enabling More ❯
We'd Like to See: Languages/Scripting: JavaScript Application deployment: Ansible OGC Web Services: WMS, WMTS, WCS, WFS Frameworks: AI/ML Data: PySpark, Elasticsearch, Kibana Other Skills: Apache NiFi, ApacheAccumulo, High performance computing cluster Certifications: Sec+ Who we are: Reinventing Geospatial, Inc. (RGi) is a fast-paced small business that has the environment and More ❯
Boot Experience with processing Big Data Demonstrated experience with system design and architecture Experience with Web development, HTTP, and REST services Experience with NoSQL technologies such as Elasticsearch and Accumulo Experience with CI/CD principles, concepts, best practices and tools such as Jenkins and GitLab CI Position Desired Skills: Experience with the Atlassian Tool suite (Jira, Confluence) Experience More ❯
Boot Experience with processing Big Data Demonstrated experience with system design and architecture Experience with Web development, HTTP, and REST services Experience with NoSQL technologies such as Elasticsearch and Accumulo Experience with CI/CD principles, concepts, best practices and tools such as Jenkins and GitLab CI Position Desired Skills Experience with the Atlassian Tool suite (Jira, Confluence) Experience More ❯
Pig is highly desired • Experience with Data Science • Experience with Graphic Algorithms • Experience with Machine Learning • Experience with AWS • Cloud development experience such as Hadoop, Big Data (Cloudbase/Accumulo and Big Table) as well as JSON/BSON • Experience with analytic development • Experience with Python and streaming capabilities • Experience with Software Configuration Management Tools such as JIRA, GIT More ❯
US citizenship and an active Secret clearance; will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity More ❯
US citizenship and an active Secret clearance; will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity More ❯
US citizenship and an active Secret clearance; will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity More ❯
San Antonio, Texas, United States Hybrid / WFH Options
IAMUS
during the interview process. Place of Performance: either Columbia, MD or San Antonio, TX. Both positions would involve Hybrid work. Desired Skills (Optional) Experience with NOSQL databases such as Accumulo desired CI Poly preferred Experience developing with Kubernetes environments Prior Experience supporting cyber and/or network security operations within a large enterprise, as either an analyst, engineer, architect More ❯
the position Computer Science (CS) degree or related field Experience with Java, Python, C, and Query Time Analytics (QTA) Customer GHOSTMACHINE analytic development Experience with Hadoop (Map Reduce, and Accumulo) Experience with Linux Experience with GEOINT Desired Skills: Familiarity with JIRA and Confluence. Understanding of customer analytical tools' Compensation Range: $198,094.13 - $223,094.13 _ Compensation ranges encompass a total More ❯
Herndon, Virginia, United States Hybrid / WFH Options
Vantor
REST APIs, microservices , IaaS/PaaS services. D eveloping and deploying web services. W orking with open-source resources in a government computing environment Big data technologies such as Accumulo , Spark, Hive, Hadoop, ElasticSearch Strong Linux skills and familiarity with hybrid cloud/on-prem architecture. CompTIA Security+ or comparable certification for privileged user access. Experience a s a More ❯
and DevSecOps practices. Proficient in Postgres design/optimization, distributed processing (REST APIs, microservices, IaaS/PaaS), and developing/deploying web services. Experience with big data technologies, including Accumulo, Spark, Hive, Hadoop, and ElasticSearch. Experience working with open-source resources in government computing environments; military/intelligence analyst experience or familiarity with IC PED systems is a plus. More ❯
San Antonio, Texas, United States Hybrid / WFH Options
Wyetech, LLC
Understanding of AGILE software development methodologies and use of standard software development tool suites Desired Technical Skills Security+ certification is highly desired. Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Work could possibly require More ❯
skills and familiarity with hybrid cloud/on-prem architecture, AWS, C2S, Openstack, etc. Experience with some big data technologies such as Kubernetes, Spark, Hive, and/or Hadoop, Accumulo, ElasticSearch Experience with Apache NiFi, Apache Airflow, or Kafka An adaptable and solution centric mindset that embraces technology enablers. Familiarity with common industry software tools, concepts, and More ❯
San Antonio, Texas, United States Hybrid / WFH Options
IAMUS
and ability to build interactive, insightful dashboards for monitoring ML models. Place of Performance: Hybrid work in San Antonio, TX. Desired Skills (Optional) Experience with NOSQL databases such as Accumulo is a plus Familiarity with deploying models as API's or within containerized environments (e.g., Docker, Kubernetes) to serve and monitor models in production. Experience with Large Language Models More ❯
San Antonio, Texas, United States Hybrid / WFH Options
IAMUS
Flexibility is key to accommodate any schedules changes per the customer and team in place. Preferred Requirements Security+ certification is highly desired. Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Work could possibly require More ❯
the schedule. Must be willing/able to help open/close the workspace during regular business hours as needed. Preferred Requirements Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Compensation At IAMUS Consulting More ❯
software to correct defects, adapt to new hardware, or improve performance, ensuring integration with Hadoop Distributed File System (HDFS) environments and distributed Big Data stores (e.g., HBase, CloudBase/Accumulo, Big Table). Must have the ability to develop simple data queries and complex database or data repository interfaces for existing and proposed databases, utilizing serialization formats such as More ❯
the schedule. • Must be willing/able to help open/close the workspace during regular business hours as needed. Preferred Requirements • Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. • Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Benefits More ❯
San Antonio, Texas, United States Hybrid / WFH Options
HII Mission Technologies
Flexibility is key to accommodate any schedules changes per the customer and team in place. Preferred Requirements Security+ certification is highly desired. Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Work could possibly require More ❯
/week. Must be willing/able to help open/close the workspace during regular business hours as needed. Preferred Requirements Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. We have many more More ❯
/week. Must be willing/able to help open/close the workspace during regular business hours as needed. Preferred Requirements Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. We have many more More ❯