search platforms, GPU workloads, and distributed storage (e.g., Cloudera). Experience developing algorithms with R, Python, SQL, or NoSQL. Knowledge of distributed data and computing tools such as Hadoop, Hive, Spark, MapReduce, or EMR. Hands-on experience with visualization tools like Plotly, Seaborn, or ggplot2. Security+ certification. More ❯
SCI with Full Scope Polygraph. Desired Qualifications: Experience with AWS Data Management services (Elastic Map Reduce, Lambda, Kinesis). Experience with SAFe development practices. Experience with Python, SpringBoot, Hibernate, Hive, Pig, or C++. More ❯
Big Data Store (NoSQL) such as Hbase, CloudBase/Acumulo, Big Table, etc.; Shall have demonstrated work experience with the Map Reduce programming model and technologies such as Hadoop, Hive, Pig, etc.; Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS); Shall have demonstrated work experience with Serialization such as JSON and/or BSON More ❯
desirably have knowledge of modeling techniques (logit, GLM, time series, decision trees, random forests, clustering), statistical programming languages (SAS, R, Python, Matlab) and big data tools and platforms (Hadoop, Hive, etc.). Solid academic record. Strong computer skills. Knowledge of other languages is desirable. Get-up-and-go attitude, maturity, responsibility and strong work ethic. Strong ability to learn More ❯
desirably have knowledge of modeling techniques (logit, GLM, time series, decision trees, random forests, clustering), statistical programming languages (SAS, R, Python, Matlab) and big data tools and platforms (Hadoop, Hive, etc.). Solid academic record. Strong computer skills. Knowledge of other languages is desirable. Get-up-and-go attitude, maturity, responsibility and strong work ethic. Strong ability to learn More ❯
The role also involves optimizing database architecture and performance, implementing DevSecOps practices, and building CI/CD pipelines using Python, Bash, and Terraform. Preferred candidates will have experience with Apache Spark, Apache Nifi, data governance, and ETL standardization. Familiarity with Glue, Hive, and Iceberg or similar technologies is a plus. Tasks Performed: • Bridge communication between technical staff … data between systems, and optimize queries. • Plan and execute large-scale data migrations. • Improve database performance through architecture and tuning. • Create and maintain data flows using ETL tools like Apache Nifi. • Manage infrastructure as code using Python, Bash, and Terraform. • Integrate security into development and deployment workflows. • Build and support automated CI/CD pipelines. Education, Experience and Qualifications … mapping between SQL databases. • Demonstrated experience in large-scale data migration efforts. • Demonstrated experience with database architecture, performance design methodologies, and system-tuning recommendations. Preference for familiarity with Glue, Hive, and Iceberg or similar. • Demonstrated experience with Python, Bash, and Terraform. • Demonstrated experience with DevSecOps solutions and tools. • Demonstrated experience implementing CI/CD pipelines using industry standard process. More ❯
support highly distributed, massively parallel computation needssuch as Hbase, CloudBase/Acumulo, Big Table. Shall have demonstrated work experience with the Map Reduce programming model and technologiessuch as Hadoop, Hive, Pig. Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS). Shall have demonstrated work experience with Serialization such as JSON and/or BSON. Shall More ❯
work experience with: o Distributed scalable Big Data Store (NoSQL) such as Hbase, CloudBase/Acumulo, Big Table, etc. o Map Reduce programming model and technologies such as Hadoop, Hive, Pig, etc. o Hadoop Distributed File System (HDFS) o Serialization such as JSON and/or BSON • 4 years of SWE experience may be substituted for a bachelor's More ❯
also have Systems integration background or experience Experience of developing the Finance Data Strategy for large financial institutions, developing future state architecture Delivery experience in Big Data technologies and Apache ecosystem technologies such as Spark, Kafka, Hive etc and have experience building end to end data pipelines using on-premise or cloud-based data platforms. Hands-on experience More ❯
herndon, virginia, united states Hybrid / WFH Options
Maxar Technologies
User Interface, Databases Strong Linux skills and familiarity with hybrid cloud/on-prem architecture, AWS, C2S, Openstack, etc. Experience with some big data technologies such as Kubernetes, Spark, Hive, and/or Hadoop, Accumulo, ElasticSearch Experience with Apache NiFi, Apache Airflow, or Kafka An adaptable and solution centric mindset that embraces technology enablers. Familiarity with common More ❯
Spark - Must have Scala - Must Have hands on coding Hive & SQL - Must Have Note: Please screen the profile before interview. At least Candidate should know Scala coding language. Pyspark profile will not help here. Interview includes coding test. Job Description: = Scala/Spark Good Big Data resource with the below Skillset: § Spark § Scala § Hive/HDFS/HQL … Linux Based Hadoop Ecosystem (HDFS, Impala, Hive, HBase, etc.) Experience in Big data technologies, Real Time data processing platform (Spark Streaming) experience would be an advantage. Consistently demonstrates clear and concise written and verbal communication A history of delivering against agreed objectives Ability to multi-task and work under pressure Demonstrated problem solving and decision-making skills Excellent analytical More ❯
Modelling : Physical, logical, conceptual models; data flow diagrams; ontologies; UML/Visio/Sparx Data Standards : Technical specs, code assurance, championing interoperability Metadata Management : Data catalogues, repositories; tools like Apache Atlas, Hive Metastore, AWS Glue/Datazone Data Design : Data lakes, warehouses, lakehouses, pipelines, meshes, marketplaces If you're passionate about shaping data strategy and architecture in a More ❯