Python, Go, Julia etc.) •Experience with Amazon Web Services (S3, EKS, ECR, EMR, etc.) •Experience with containers and orchestration (e.g. Docker, Kubernetes) •Experience with Big Data processing technologies (Spark, Hadoop, Flink etc) •Experience with interactive notebooks (e.g. JupyterHub, Databricks) •Experience with Git Ops style automation •Experience with ix (e.g, Linux, BSD, etc.) tooling and scripting •Participated in projects that More ❯
with Spark. Experience building, maintaining, and debugging DBT pipelines. Strong proficiency in developing, monitoring, and debugging ETL jobs. Deep understanding of SQL and experience with Databricks, Snowflake, BigQuery, Azure, Hadoop, or CDP environments. Hands-on technical support experience, including escalation management and adherence to SLAs. Familiarity with CI/CD technologies and version control systems like Git. Expertise in More ❯
with Spark. Experience building, maintaining, and debugging DBT pipelines. Strong proficiency in developing, monitoring, and debugging ETL jobs. Deep understanding of SQL and experience with Databricks, Snowflake, BigQuery, Azure, Hadoop, or CDP environments. Hands-on technical support experience, including escalation management and adherence to SLAs. Familiarity with CI/CD technologies and version control systems like Git. Expertise in More ❯
really make yourapplication stand out: Implementationexperience with Machine Learning models and applications Knowledgeof cloud-based Machine Learning engines (AWS, Azure, Google, etc.) Experiencewith large scale data processing tools (Spark, Hadoop, etc.) Abilityto query and program databases (SQL, No SQL) Experiencewith distributed ML frameworks (TensorFlow, PyTorch, etc.) Familiaritywith collaborative software tools (Git, Jira, etc.) Experiencewith user interface libraries/applications More ❯
Tableau) Machine Learning Fundamentals (e.g., Supervised, Unsupervised Learning) Machine Learning Algorithms (e.g., Regression, Classification, Clustering, Decision Trees, SVMs, Neural Networks) Model Evaluation and Validation Big Data Technologies (e.g., Spark, Hadoop - conceptual understanding) Database Querying (e.g., SQL) Cloud-based Data Platforms (e.g., AWS Sagemaker, Google AI Platform, Azure ML) Ethics in Data Science and AI Person Specification: Experience supporting data More ❯
utilising strong communication and stakeholder management skills when engaging with customers Significant experience of coding in Python and Scala or Java Experience with big data processing tools such as Hadoop or Spark Cloud experience; GCP specifically in this case, including services such as Cloud Run, Cloud Functions, BigQuery, GCS, Secret Manager, Vertex AI etc. Experience with Terraform Prior experience More ❯
Azure Data Factory, Azure Functions, and Synapse Analytics. Proficient in Python and advanced SQL, including query tuning and optimisation. Hands-on experience with big data tools such as Spark, Hadoop, and Kafka. Familiarity with CI/CD pipelines, version control, and deployment automation. Experience using Infrastructure as Code tools like Terraform. Solid understanding of Azure-based networking and cloud More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Chi Square Analytics
machine learning techniques, including clustering, Bayesian methods, reinforcement learning, and scalable modelling Technical fluency in tools and platforms like Python or R, cloud environments, and big data infrastructure (e.g., Hadoop, Kafka, SQL) Exceptional stakeholder communication-able to translate complex technical ideas into strategic insights for executive audiences Bonus: Experience in industries like gaming, entertainment or technology would add value More ❯
machine learning techniques, including clustering, Bayesian methods, reinforcement learning, and scalable modelling Technical fluency in tools and platforms like Python or R, cloud environments, and big data infrastructure (e.g., Hadoop, Kafka, SQL) Exceptional stakeholder communication-able to translate complex technical ideas into strategic insights for executive audiences Bonus: Experience in industries like gaming, entertainment or technology would add value More ❯
environment) Clojure or similar languages (e.g. Java or Python) Software collaboration and revision control (e.g. Git or SVN) Desired skills and experiences: ElasticSearch/Kibana Cloud computing (e.g. AWS) Hadoop/Spark etc. Graph Databases Educational level: Master Degree Tagged as: Clustering , Data Mining , Industry , Information Retrieval , Master Degree , Sentiment Analysis , United Kingdom More ❯
expertise Demonstrable experience of utilising strong communication and stakeholder management skills when engaging with customers Significant experience of coding (ideally Python) Experience with big data processing tools such as Hadoop or Spark Cloud experience; GCP specifically in this case, including services such as Cloud Run, Cloud Functions, BigQuery, GCS, Secret Manager, Vertex AI etc. Experience with Terraform Prior experience More ❯
techniques Experience delivering analytics projects across different business areas, demonstrating strong business acumen Proficiency in probability and statistics Advanced SQL skills, querying tools (e.g. qubole/big query/hadoop), and concepts such as Store Procedures, Cursors and Temp tables. Proficiency in Python (knowledge of R is also welcome) Expertise in Tableau Strong data visualization skills for communicating results More ❯
with ability to think strategically about business, product, and technical challenges in an enterprise environment - Extensive hands-on experience with data platform technologies, including at least three of: Spark, Hadoop ecosystem, orchestration frameworks, MPP databases, NoSQL, streaming technologies, data catalogs, BI and visualization tools - Proficiency in at least one programming language (e.g., Python, Java, Scala), infrastructure as code, cloud More ❯
of platform components. Big Data Architecture: Build and maintain big data architectures and data pipelines to efficiently process large volumes of geospatial and sensor data. Leverage technologies such as Hadoop, Apache Spark, and Kafka to ensure scalability, fault tolerance, and speed. Geospatial Data Integration: Develop systems that integrate geospatial data from a variety of sources (e.g., satellite imagery, remote … data-driven applications. Familiarity with geospatial data formats (e.g., GeoJSON, Shapefiles, KML) and tools (e.g., PostGIS, GDAL, GeoServer). Technical Skills: Expertise in big data frameworks and technologies (e.g., Hadoop, Spark, Kafka, Flink) for processing large datasets. Proficiency in programming languages such as Python, Java, or Scala, with a focus on big data frameworks and APIs. Experience with cloud More ❯
e.g., SQL, PL/SQL, DDL, MDX, HiveQL, SparkSQL, Scala) - Experience with one or more scripting language (e.g., Python, KornShell) PREFERRED QUALIFICATIONS - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR - Experience with any ETL tool like, Informatica, ODI, SSIS, BODI, Datastage, etc. Our inclusive culture empowers Amazonians to deliver the best results for our customers. If More ❯
challenges. You are proficient in Python, with experience using PySpark and ML libraries such as scikit-learn, TensorFlow, or Keras . You are familiar with big data technologies (e.g., Hadoop, Spark), cloud platforms (AWS, GCP), and can effectively communicate technical concepts to non-technical stakeholders. Accommodation requests If you need assistance with any part of the application or recruiting More ❯
with some of the brightest technical minds in the industry today. BASIC QUALIFICATIONS - 10+ years of technical specialist, design and architecture experience - 10+ years of database (eg. SQL, NoSQL, Hadoop, Spark, Kafka, Kinesis) experience - 10+ years of consulting, design and implementation of serverless distributed solutions experience - Australian citizen with ability to obtain security clearance. PREFERRED QUALIFICATIONS - AWS Professional level More ❯
native tech stack in designing and building data & AI solutions Experience with data modeling, ETL processes, and data warehousing Knowledge of big data tools and frameworks such as Spark, Hadoop, or Kafka More ❯
real-time intelligence, BI, Copilot), Azure Databricks, Purview Data Governance, and Azure Databases: SQL DB, Cosmos DB, PostgreSQL. Maintain and grow expertise in on-prem EDW (Teradata, Netezza, Exadata), Hadoop & BI solutions. Represent Microsoft through thought leadership in cloud Database & Analytics communities and customer forums. Qualifications Proven technical pre-sales or technical consulting experience. OR Bachelor's Degree in More ❯
Agile working practices CI/CD tooling Scripting experience (Python, Perl, Bash, etc.) ELK (Elastic stack) JavaScript Cypress Linux experience Search engine technology (e.g., Elasticsearch) Big Data Technology experience (Hadoop, Spark, Kafka, etc.) Microservice and cloud native architecture Desirable skills Able to demonstrate experience of troubleshooting and diagnosis of technical issues. Able to demonstrate excellent team-working skills. Strong More ❯
senior leadership PREFERRED QUALIFICATIONS - Experience working as a BIE in a technology company - Experience with AWS technologies - Experience using Cloud Storage and Computing technologies such as AWS Redshift, S3, Hadoop, etc. Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application More ❯
Previous experience as a Data Engineer (3-5 years); Deep expertise in designing and implementing solutions on Google Cloud; Strong interpersonal and stakeholder management skills; In-depth knowledge of Hadoop, Spark, and similar frameworks; In-depth knowledge of programming languages including Java; Expert in cloud-native technologies, IaC, and Docker tools; Excellent project management skills; Excellent communication skills; Proactivity More ❯
models in production and adjusting model thresholds to improve performance Experience designing, running, and analyzing complex experiments or leveraging causal inference designs Experience with distributed tools such as Spark, Hadoop, etc. A PhD or MS in a quantitative field (e.g., Statistics, Engineering, Mathematics, Economics, Quantitative Finance, Sciences, Operations Research) Office-assigned Stripes spend at least 50% of the time More ❯
internal customer facing, complex and large scale project management experience - 5+ years of continuous integration and continuous delivery (CI/CD) experience - 5+ years of database (eg. SQL, NoSQL, Hadoop, Spark, Kafka, Kinesis) experience - 5+ years of consulting, design and implementation of serverless distributed solutions experience - 3+ years of cloud based solution (AWS or equivalent), system, network and operating More ❯
NLP technologies Proficiency in Python and/or Scala; experience with ML libraries such as TensorFlow, PyTorch, HuggingFace, or scikit-learn Experience with Databricks, distributed data systems (e.g., Spark, Hadoop), and cloud platforms (AWS, GCP, or Azure) Ability to thrive in ambiguous environments, working closely with cross-functional teams to define and deliver impactful solutions Strong communication skills with More ❯