and data pipeline orchestration tools such as Apache Airflow and Nifi. Experience with large scale/Big Data technologies, such as Hadoop, Spark, Hive, Impala, PrestoDb, Kafka. Experience with workflow orchestration tools like Apache Airflow. Experience with containerisation using Docker and deployment on Kubernetes. Experience with NoSQL and graph More ❯
data analytics on AWS platforms. Experience in writing efficient SQLs, implementing complex ETL transformations on big data platforms. Experience in Big Data technologies (Spark, Impala, Hive, Redshift, Kafka, etc.). Experience in data quality testing; adept at writing test cases and scripts, presenting and resolving data issues. Experience with More ❯
is key to accommodate any schedules changes per the customer. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers EKS, Diode, CI/CD, and Terraform are a plus Work could possibly require some on-call work. More ❯
San Antonio, Texas, United States Hybrid / WFH Options
IAMUS
to accommodate schedule changes based on customer and team needs. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. Compensation At IAMUS Consulting, we More ❯
Flexibility is essential to adapt to schedule changes if needed. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers EKS, Diode, CI/CD, and Terraform are a plus Work could possibly require some on-call work. More ❯
adapt to schedule changes as needed. Preferred Requirements Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Work could possibly require some on-call work. More ❯
during regular business hours as needed. Preferred Requirements Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. We have many more additional great benefits/ More ❯
San Antonio, Texas, United States Hybrid / WFH Options
Syndicus NACON
to accommodate schedule changes based on customer and team needs. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. Benefits More ❯
Flexibility is essential to adapt to schedule changes if needed. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers EKS, Diode, CI/CD, and Terraform are a plus Work could possibly require some on-call work. More ❯
is key to accommodate any schedules changes per the customer. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. We have many more additional More ❯
Flexibility is essential to adapt to schedule changes if needed. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers EKS, Diode, CI/CD, and Terraform are a plus Work could possibly require some on-call work. More ❯
San Antonio, Texas, United States Hybrid / WFH Options
Enlighten, an HII - Mission Technologies Company
any schedules changes per the customer and team in place. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. We have many more additional More ❯
Exposure and hands on in Microservices, Distributed Cache (REDIS, Couchbase) and Cloud technologies • Good to have knowledge and experience in Big data - HBASE and Impala concepts. • Experienced with XML parsing (including schemas), JSON and third-party libraries like Gauva, lombok. • Well versed with design standards & frameworks; experience in working More ❯
is essential to adapt to schedule changes if needed. Desired Skills (Optional) Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. About The DarkStar Group Our More ❯
is essential to adapt to schedule changes if needed. Desired Skills (Optional) Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus Work could possibly require some on-call work. About The DarkStar Group Our More ❯
Hanover, Maryland, United States Hybrid / WFH Options
Enlighten, an HII - Mission Technologies Company
Flexibility is essential to accommodate any changes in the schedule. Preferred Requirements Experience with big data technologies like: Hadoop, Spark, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers EKS, Diode, CI/CD, and Terraform are a plus We have many more additional great benefits/ More ❯
Alexandria, Virginia, United States Hybrid / WFH Options
Metronome LLC
/close the workspace during regular business hours as needed Desired Skills Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes are a plus All candidates will be required to be on-site at a manager identified More ❯
Extensive knowledge of popular database providers such as SQL Server, PostgreSQL, Teradata and others. • Proficiency in technologies in the Apache Hadoop ecosystem, especially Hive, Impala and Ranger • Experience working with open file and table formats such Parquet, AVRO, ORC, Iceberg and Delta Lake • Extensive knowledge of automation and software More ❯
have a DoD 8140/8570 compliance certification (i.e. Security+ certification) Preferred Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes More ❯
and technologies such as JSON, XML, XSLT, JDBC, SOAP and REST. Experience with Cloud-based data analysis tools including Hadoop and Mahout, Acumulo, Hive, Impala, Pig, and similar. Experience with visual analytic tools like Microsoft Pivot, Palantir, or Visual Analytics. Experience with open-source textual processing such as Lucene More ❯
Data Mining, Classical Machine Learning, Deep Learning, NLP and Computer Vision. Experience with Large Scale/Big Data technology, such as Hadoop, Spark, Hive, Impala, PrestoDb. Hands-on capability developing ML models using open-source frameworks in Python and R and applying them on real client use cases. Proficient More ❯
reviews, testing, and debugging. Ensure performance, quality, and responsiveness of applications. Key Requirements and Technology Experience: Key Skills:Python, pandas, NumPy, Hadoop ecosystem (Hive, Impala, PySpark, spark), SQL queries. Strong hands-on experience with Python and related . libraries: pandas , NumPy , PySpark , Spark . Solid understanding and experience with … the Hadoop ecosystem . Proficiency with Hive , Impala , and writing complex SQL queries . Experience using Jupyter Notebook , Notepad , or similar tools for development and documentation . Exposure to Apache Kafka for real-time data processing . Bachelor's or Master's degree in Computer Science, Engineering, or related More ❯
and AI workloads using tools like Jupyter, Spacy, Transformers, and NLTK. Big Data Platforms: Utilize big data NoSQL engines and platforms such as Hive, Impala, and Elasticsearch for data storage and processing BI and Visualization: Implement and support business intelligence and visualization tools like Tableau, Kibana, and PowerBI to … science, machine learning, and AI tools such as Jupyter, Spacy, Transformers, and NLTK. Experience with big data NoSQL engines/platforms such as Hive, Impala, and Elasticsearch. Proficiency with business intelligence and visualization tools like Tableau, Kibana, and PowerBI. Excellent communication and collaboration skills. Preferred Qualifications: Certification in AWS More ❯
and AI workloads using tools like Jupyter, Spacy, Transformers, and NLTK. Big Data Platforms: Utilize big data NoSQL engines and platforms such as Hive, Impala, and Elasticsearch for data storage and processing BI and Visualization: Implement and support business intelligence and visualization tools like Tableau, Kibana, and PowerBI to … science, machine learning, and AI tools such as Jupyter, Spacy, Transformers, and NLTK. Experience with big data NoSQL engines/platforms such as Hive, Impala, and Elasticsearch. Proficiency with business intelligence and visualization tools like Tableau, Kibana, and PowerBI. Excellent communication and collaboration skills. Preferred Qualifications: Certification in AWS More ❯
and maintain Dash web application with user-friendly interface for workflow processing, data visualization, exploration, and efficient reporting. Design and implement relational databases in Impala to effectively store and manage data. Develop optimal schemas for Impala tables based on query patterns and data characteristics. Integrate Dash applications with … Impala to efficiently query and process large data sets. Implement and manage Oozie job schedulers for maintaining ETL processes to efficiently load, transform, and distribute daily data. Employ agile development practices to develop effective business solutions based on the business needs. Required Skills Education & Experience: Master's or higher … manipulation and analysis using libraries such as Pandas, NumPy, and SQLAlchemy. Extensive experience with Dash framework for building web applications. In-depth knowledge of Impala or other SQL-on-Hadoop query engines. Understanding of web development concepts (HTML, CSS, JavaScript). Proficiency in data visualization libraries (Plotly, Seaborn). More ❯