in at least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or Spark Streaming) 4+ years of experience in open source frameworks At this time More ❯
in at least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or Spark Streaming) 4+ years of experience in open source frameworks At this time More ❯
in at least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or Spark Streaming) 4+ years of experience in open source frameworks At this time More ❯
in at least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or Spark Streaming) 4+ years of experience in open source frameworks At this time More ❯
the IC. Expert proficiency in Python (or similar languages) and experience with data science libraries (TensorFlow, PyTorch, Pandas, NumPy). Strong experience with big data processing tools (e.g., Spark, Hadoop, AWS or Azure cloud platforms). Expertise in working with geospatial data formats (e.g., GeoTIFF, Shapefiles, WMS, WFS) and spatial libraries (e.g., GeoPandas, Rasterio, GDAL). Advance experience in More ❯
the IC. Expert proficiency in Python (or similar languages) and experience with data science libraries (TensorFlow, PyTorch, Pandas, NumPy). Strong experience with big data processing tools (e.g., Spark, Hadoop, AWS or Azure cloud platforms). Expertise in working with geospatial data formats (e.g., GeoTIFF, Shapefiles, WMS, WFS) and spatial libraries (e.g., GeoPandas, Rasterio, GDAL). Advance experience in More ❯
in at least one of the following: Java, Scala, Python, Go, Javascript/TypeScript, Angular/React.js, or Node.js 2+ years of experience working with big data technologies (e.g. Hadoop, Spark, Presto) 2+ years of experience working on streaming data applications (e.g. Kafka, Kinesis, Flink, or Spark Streaming) 4+ years of experience in open source frameworks At this time More ❯
San Antonio, Texas, United States Hybrid / WFH Options
Wyetech, LLC
skills. Understanding of AGILE software development methodologies and use of standard software development tool suites Desired Technical Skills Security+ certification is highly desired. Experience with big data technologies like: Hadoop, Accumulo, Ceph, Spark, NiFi, Kafka, PostgreSQL, ElasticSearch, Hive, Drill, Impala, Trino, Presto, etc. Experience with containers, EKS, Diode, CI/CD, and Terraform are a plus. Work could possibly More ❯
hands-on experience in programming and software development using Java, JavaScript, or Python. Demonstrated hands on experience working with PostgreSQL and Apache NiFi. Demonstrated hands-on experience working with Hadoop, Apache Spark and their related ecosystems. A candidate must be a US Citizen and requires an active/current TS/SCI with Polygraph clearance. Salary Range More ❯
and databases. Strong Linux skills with experience in hybrid cloud/on-prem architectures (AWS, C2S, OpenStack, etc.). Experience with big data technologies such as Kubernetes, Spark, Hive, Hadoop, Accumulo, and ElasticSearch. Experience with workflow and streaming tools such as Apache NiFi, Apache Airflow, or Kafka. Knowledge of common industry software tools, DevSecOps practices, and working with open More ❯
basis. Skills Required • 10-12 years of technology experience • Solid working experience in Python • Proven experience as a Machine Learning Engineer or similar role • Good knowledge of Data Lake, Hadoop, Dataiku, Scala and Apache Spark • Working knowledge of SQL and No SQL DBs • Deep knowledge of math, probability, statistics and algorithms • Working knowledge of machine learning frameworks (like Keras More ❯
automation testing of etl (extract, transform, load) processes and data warehousing. 3. Strong understanding of sql for data querying and validation. 4. Knowledge of big data technologies such as hadoop, spark, or kafka is a plus. 5. Familiarity with scripting languages like python, java, or shell scripting. 6. Excellent analytical and problem-solving skills with a keen attention to More ❯
automation testing of etl (extract, transform, load) processes and data warehousing. 3. Strong understanding of sql for data querying and validation. 4. Knowledge of big data technologies such as hadoop, spark, or kafka is a plus. 5. Familiarity with scripting languages like python, java, or shell scripting. 6. Excellent analytical and problem-solving skills with a keen attention to More ❯
automation testing of etl (extract, transform, load) processes and data warehousing. 3. Strong understanding of sql for data querying and validation. 4. Knowledge of big data technologies such as hadoop, spark, or kafka is a plus. 5. Familiarity with scripting languages like python, java, or shell scripting. 6. Excellent analytical and problem-solving skills with a keen attention to More ❯
automation testing of etl (extract, transform, load) processes and data warehousing. 3. Strong understanding of sql for data querying and validation. 4. Knowledge of big data technologies such as hadoop, spark, or kafka is a plus. 5. Familiarity with scripting languages like python, java, or shell scripting. 6. Excellent analytical and problem-solving skills with a keen attention to More ❯
london (city of london), south east england, united kingdom
HCLTech
automation testing of etl (extract, transform, load) processes and data warehousing. 3. Strong understanding of sql for data querying and validation. 4. Knowledge of big data technologies such as hadoop, spark, or kafka is a plus. 5. Familiarity with scripting languages like python, java, or shell scripting. 6. Excellent analytical and problem-solving skills with a keen attention to More ❯
leadership experience Proficiency in modern data platforms (Databricks, Snowflake, Kafka), container orchestration (Kubernetes/OpenShift), and multi-cloud deployments across AWS, Azure, GCP Advanced knowledge of Big Data ecosystems (Hadoop/Hive/Spark), data lakehouse architectures, mesh topologies, and real-time streaming platforms Strong Unix/Linux skills, database connectivity (JDBC/ODBC), authentication systems (LDAP, Active Directory More ❯
leadership experience Proficiency in modern data platforms (Databricks, Snowflake, Kafka), container orchestration (Kubernetes/OpenShift), and multi-cloud deployments across AWS, Azure, GCP Advanced knowledge of Big Data ecosystems (Hadoop/Hive/Spark), data lakehouse architectures, mesh topologies, and real-time streaming platforms Strong Unix/Linux skills, database connectivity (JDBC/ODBC), authentication systems (LDAP, Active Directory More ❯
leadership experience Proficiency in modern data platforms (Databricks, Snowflake, Kafka), container orchestration (Kubernetes/OpenShift), and multi-cloud deployments across AWS, Azure, GCP Advanced knowledge of Big Data ecosystems (Hadoop/Hive/Spark), data lakehouse architectures, mesh topologies, and real-time streaming platforms Strong Unix/Linux skills, database connectivity (JDBC/ODBC), authentication systems (LDAP, Active Directory More ❯
years, associate's with 10 years, bachelor's with 8 years, master's with 6 years, or PhD with 4 years Deep expertise in big data platforms (e.g., Hadoop, Spark, Kafka) and multi-cloud environments (AWS, Azure, GCP) Experience with machine learning frameworks (e.g., TensorFlow, Scikit-learn, PyTorch) Strong programming skills in Python, Java, or Scala Familiarity with data More ❯
london (city of london), south east england, united kingdom
Computappoint
leadership experience Proficiency in modern data platforms (Databricks, Snowflake, Kafka), container orchestration (Kubernetes/OpenShift), and multi-cloud deployments across AWS, Azure, GCP Advanced knowledge of Big Data ecosystems (Hadoop/Hive/Spark), data lakehouse architectures, mesh topologies, and real-time streaming platforms Strong Unix/Linux skills, database connectivity (JDBC/ODBC), authentication systems (LDAP, Active Directory More ❯
leadership experience Proficiency in modern data platforms (Databricks, Snowflake, Kafka), container orchestration (Kubernetes/OpenShift), and multi-cloud deployments across AWS, Azure, GCP Advanced knowledge of Big Data ecosystems (Hadoop/Hive/Spark), data lakehouse architectures, mesh topologies, and real-time streaming platforms Strong Unix/Linux skills, database connectivity (JDBC/ODBC), authentication systems (LDAP, Active Directory More ❯
Proficiency in data science languages and tools (e.g., Python, R, SQL, Jupyter, Pandas, Scikit-learn) Experience with machine learning frameworks (e.g., TensorFlow, PyTorch) and big data platforms (e.g., Spark, Hadoop) Strong background in statistics, data modeling, and algorithm development Ability to explain complex data findings to technical and non-technical stakeholders Experience supporting national security or defense data programs More ❯
relational database concepts • Experience with statistical programming languages such as R, Python, MATLAB • Knowledge of geospatial analysis concepts and tools such as ArcGIS • Familiarity with big data technologies like Hadoop, Spark, NoSQL databases • Excellent problem-solving skills and ability to develop innovative solutions • Strong verbal and written communication skills Desired Qualifications 5+ years of experience in data science, preferably More ❯
Support for platform components like document indexing, GPU workloads, and distributed storage (e.g., Cloudera) -Algorithm development using R, Python, SQL, or NoSQL -Experience with distributed computing tools (e.g., MapReduce, Hadoop, Hive, EMR, Spark, Gurobi, MySQL) -Visualization experience with Plotly, Seaborn, or ggplot2 More ❯