and ETL - informatica - Experience in SQL and database management systems - Knowledge of data modelling , warehousing concepts , and ETL processes - Experience with big data technologies and frameworks such as Hadoop, Hive, Spark. Programming experience in Python or Scala. - Demonstrated analytical and problem-solving skills. - Familiarity with cloud platforms (e.g Azure , AWS ) and their data related services - Proactive and detail oriented More ❯
data architectures, Lambda type architectures - Proficiency in writing and optimizing SQL - Knowledge of AWS services including S3, Redshift, EMR, Kinesis and RDS. - Experience with Open Source Data Technologies (Hadoop, Hive, Hbase, Pig, Spark, etc.) - Ability to write code in Python, Ruby, Scala or other platform-related Big data technology - Knowledge of professional software engineering practices & best practices for the More ❯
Experience with PostgreSQL, ElasticSearch, MongoDB, and graph databases. Experience with GenAI enabled daily workflows (coding, testing, analytics). Experience with one or more SQL-on-Hadoop technology (Spark SQL, Hive, Impala, Presto, etc.) Experience with short-release cycles and the full software lifecycle Experience with Agile development methodology (e.g., Scrum) Current Security+ certification (or must obtain within 30 days More ❯
machine learning libraries/packages (sklearn, TensorFlow, PyTorch, statsmodels, etc.). Experience in multiple tools/language/frameworks within the Big Data & cloud ecosystem (Hadoop, MongoDB, Neo4j, Spark, Hive, HBase, Cassandra, etc.). Demonstrated experience of managing and mentoring teams of data scientists, ML engineers, and data engineers on the execution of specific business use cases for AI More ❯
will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity/Affirmative Action employer. All qualified applicants More ❯
will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity/Affirmative Action employer. All qualified applicants More ❯
will also accept TS/SCI or TS/SCI with CI Polygraph Desired Experience: Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Work could possibly require some on-call work. The Swift Group and Subsidiaries are an Equal Opportunity/Affirmative Action employer. All qualified applicants More ❯
Charlotte, North Carolina, United States Hybrid / WFH Options
Collabera
Job Title: Hadoop Developer Location: Charlotte NC (Hybrid) Duration: 12+ Months Contract Pay range -$58 to $62 Must Haves: Strong hands-on experience with Hadoop ecosystem tools (HDFS, Hive, Sqoop, Oozie, Pig, Spark). Proficiency in Java, Python, or Scala for developing Hadoop applications. Solid knowledge of SQL and experience with relational databases (Oracle, MySQL, SQL Server, etc.). More ❯
data analytic frameworks and pipelines using databases and tools such as (but not limited to) NoSQL, SQL, NiFi, Kafka, HDInsight, MongoDB, Cassandra, Neo4j, GraphDB, OrientDB, Spark, Flink, Hadoop, Kafka, Hive, and others. • Apply distributed systems concepts and principles such as consistency and availability, liveness and safety, durability, reliability, fault-tolerance, consensus algorithms. • Administrate cloud computing and CI/CD More ❯
to help open/close the workspace during regular business hours as needed. Preferred Requirements • Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. • Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Benefits $152,000-$198,000 salary per year, depending on experience. More ❯
Title: GCP -Data Engineer Location: Philadelphia PA (Can submit who are willing to relocate) GCP Data Engineer - GCP Dataflow and Apache Beam (Key skills) Primary Skills- PySpark, Spark, Python, Big Data, GCP, Apache Beam, Dataflow, Airflow, Kafka and BigQuery GFO, Google Analytics Javascript is Must Strong Experience with Dataflow and BigQuery A person should have leading the team … Platforms (preferably GCP) provided Big Data technologies Hands-on experience with real-time streaming processing as well as high volume batch processing, and skilled in Advanced SQL, GCP BigQuery, Apache Kafka, Data-Lakes, etc. Hands-on Experience in Big Data technologies - Hadoop, Hive, and Spark, and an enterprise-scale Customer Data Platform (CDP) Experience in at least one More ❯
MySQL, RDS) US citizenship and an active TS/SCI with Full Scope Polygraph security clearance required Desired Experience: Experience with distributed databases and streaming tools (Hadoop, Spark, Yarn, Hive, Trino) Experience with Remote Desktop Protocol (RDP) technologies Experience with data access control, specifically Role-Based Access Control (RBAC) and Attribute-Based Access Control (ABAC) Familiarity with data science More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming More ❯
be advantageous if you had Experience working with relational and experience with NoSQL databases (such as tuning and optimising complex queries for highly scalable systems) and query languages (specifically Hive/SparkSQL and ANSI SQL).Experience building large scale Spark 3.x applications & data pipelines, ideally with Batch processing running on Hadoop clusters. If you had experience with messaging queues More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming More ❯
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming More ❯
london (city of london), south east england, united kingdom
HCLTech
modelling – Dimensional & transactional modelling using RDBMS, NO-SQL and Big Data technologies. Data visualization – Tools like Tableau Big data – Hadoop eco-system, Distributions like Cloudera/Hortonworks, Pig and HIVE Data processing frameworks – Spark & Spark streaming More ❯
Demonstrated experience in micro service architecture using Spring Framework, Spring Boot, Tomcat, AWS, Docker Container or Kubernetes solutions. 5. Demonstrated experience in big data solutions (Hadoop Ecosystem, MapReduce, Pig, Hive, DataStax, etc.) in support of a screening and vetting mission. More ❯
development of database solution in both traditional as well as columnar databases. Experience in data integration with ETL techniques and frameworks Experience in Big Data querying tools, such as Hive, Impala and Spark SQL Work Schedule: MULTIPLE SHIFTS AVAILABLE Work Type: Full Time More ❯
such as bash, Python, or Go Must have a DoD 8140/8570 compliance certification (i.e. Security+ certification) Preferred Experience with big data technologies like: Hadoop, Spark, MongoDB, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers and Kubernetes More ❯
search platforms, GPU workloads, and distributed storage (e.g., Cloudera). Experience developing algorithms with R, Python, SQL, or NoSQL. Knowledge of distributed data and computing tools such as Hadoop, Hive, Spark, MapReduce, or EMR. Hands-on experience with visualization tools like Plotly, Seaborn, or ggplot2. Security+ certification. More ❯
SCI with Full Scope Polygraph. Desired Qualifications: Experience with AWS Data Management services (Elastic Map Reduce, Lambda, Kinesis). Experience with SAFe development practices. Experience with Python, SpringBoot, Hibernate, Hive, Pig, or C++. More ❯
Big Data Store (NoSQL) such as Hbase, CloudBase/Acumulo, Big Table, etc.; Shall have demonstrated work experience with the Map Reduce programming model and technologies such as Hadoop, Hive, Pig, etc.; Shall have demonstrated work experience with the Hadoop Distributed File System (HDFS); Shall have demonstrated work experience with Serialization such as JSON and/or BSON More ❯
desirably have knowledge of modeling techniques (logit, GLM, time series, decision trees, random forests, clustering), statistical programming languages (SAS, R, Python, Matlab) and big data tools and platforms (Hadoop, Hive, etc.). Solid academic record. Strong computer skills. Knowledge of other languages is desirable. Get-up-and-go attitude, maturity, responsibility and strong work ethic. Strong ability to learn More ❯