understanding of Python and the machine learning ecosystem in Python (Numpy, Pandas, Scikit-learn, LightGBM, PyTorch)Knowledge of SQL and experience with relational databasesAgile, action-oriente Nice to have Apache SparkExperience working in cloud platforms (AWS, GCP, Microsoft Azure)Relevant knowledge or experience in the gaming industry #J-18808-Ljbffr More ❯
Science, Computer Science, or a related field. 5+ years of experience in data engineering and data quality. Strong proficiency in Python/Java, SQL, and data processing frameworks including Apache Spark. Knowledge of machine learning and its data requirements. Attention to detail and a strong commitment to data integrity. Excellent problem-solving skills and ability to work in a More ❯
Science, Computer Science, or a related field. 5+ years of experience in data engineering and data quality. Strong proficiency in Python/Java, SQL, and data processing frameworks including Apache Spark. Knowledge of machine learning and its data requirements. Attention to detail and a strong commitment to data integrity. Excellent problem-solving skills and ability to work in a More ❯
Science, Computer Science, or a related field. 5+ years of experience in data engineering and data quality. Strong proficiency in Python/Java, SQL, and data processing frameworks including Apache Spark. Knowledge of machine learning and its data requirements. Attention to detail and a strong commitment to data integrity. Excellent problem-solving skills and ability to work in a More ❯
Computer Science, Engineering, or a related field, or equivalent industry experience. Preferred Qualifications Experience or interest in mentoring junior engineers. Familiarity with data-centric workflows and pipeline orchestration (e.g., Apache Airflow). Proficiency in data validation, anomaly detection, or debugging using tools like Pandas, Polars, or data.table/R. Experience working with AWS or other cloud platforms. Knowledge of More ❯
with a focus on delivering and operating large scale data processing systems. Has successfully led data platform initiatives. A good understanding of data processing technologies and tools such as Apache Spark, Data Lake, Data Warehousing and SQL Databases. Proficiency in programming languages such as Python and CICD techniques to efficiently deliver change in a consistent, controlled, and high-quality More ❯
Leverage full-stack technologies including Java, JavaScript, TypeScript, React, APIs, MongoDB, Elastic Search, DMN, BPMN, and Kubernetes. Utilize data-streaming technologies such as Kafka CDC, Kafka topics, EMS, and Apache Flink. Innovate and incubate new ideas. Work on a broad range of problems involving large data sets, real-time processing, messaging, workflows, and UI/UX. Drive the full More ❯
Leverage full-stack technologies including Java, JavaScript, TypeScript, React, APIs, MongoDB, Elastic Search, DMN, BPMN, and Kubernetes. Utilize data-streaming technologies such as Kafka CDC, Kafka topics, EMS, and Apache Flink. Innovate and incubate new ideas. Work on a broad range of problems involving large data sets, real-time processing, messaging, workflows, and UI/UX. Drive the full More ❯
London, England, United Kingdom Hybrid / WFH Options
Locus Robotics
AWS or Azure) for deploying and scaling data systems. Highly desired experience with Azure, particularly Lakehouse and Eventhouse architectures. Experience with relevant infrastructure and tools including NATS, Power BI, Apache Spark/Databricks, and PySpark. Hands-on experience with data warehousing methodologies and optimization libraries (e.g., OR-Tools). Experience with log analysis, forensic debugging, and system performance tuning. More ❯
data-driven performance analysis and optimization Strong communication skills and ability to work in a team Strong analytical and problem-solving skills PREFERRED QUALIFICATIONS Experience with Kubernetes deployment architectures Apache NiFi experience Experience building trading controls within an investment bank ABOUT GOLDMAN SACHS At Goldman Sachs, we commit our people, capital, and ideas to help our clients, shareholders, and More ❯
leverage full-stack technologies including Java, JavaScript, TypeScript, React, APIs, MongoDB, Elastic Search, DMN, BPMN, and Kubernetes, utilize data-streaming technologies such as Kafka CDC, Kafka topics, EMS, and Apache Flink, innovate and incubate new ideas, work on a broad range of problems, often involving large data sets, including real-time processing, messaging, workflow, and UI/UX, drive More ❯
models in close cooperation with our data science team Experiment in your domain to improve precision, recall, or cost savings Requirements Expert skills in Java or Python Experience with Apache Spark or PySpark Experience writing software for the cloud (AWS or GCP) Speaking and writing in English enables you to take part in day-to-day conversations in the More ❯
London, England, United Kingdom Hybrid / WFH Options
Veeva Systems, Inc
recall, or cost savings Requirements Excellent communication skills, used to work in a remote environment More than 5 years of experience Expert skills in Python or Java Experience with Apache Spark Experience writing software in AWS Nice to Have Experience with Data Lakes, Lakehouses, and Warehouses (e. g. DeltaLake, Redshift) Previously worked in agile environments Experience with expert systems More ❯
London, England, United Kingdom Hybrid / WFH Options
Count Technologies Ltd
flat structure, you’ll work autonomously on business critical projects, and collaborate with others throughout our team and user base. Our tech stack includes TypeScript, Python, Node, WebAssembly, WebGL, Apache Arrow, DuckDB, Kubernetes and React. For the best possible user experience, we have developed various technologies in-house, including a custom WebGL rendering engine, data visualization library, reactive SQL More ❯
skills in Python, Java, Scala, or a similar language. Proficiency in database technologies (SQL, NoSQL, time-series databases) and data modelling. Strong understanding of data pipeline orchestration tools (e.g., Apache Airflow, Kubernetes). You thrive when working as part of a team Comfortable in a fast-paced environment Have excellent written and verbal English skills Last but not least More ❯
microservice architecture, API development. Machine Learning (ML): • Deep understanding of machine learning principles, algorithms, and techniques. • Experience with popular ML frameworks and libraries like TensorFlow, PyTorch, scikit-learn, or Apache Spark. • Proficiency in data preprocessing, feature engineering, and model evaluation. • Knowledge of ML model deployment and serving strategies, including containerization and microservices. • Familiarity with ML lifecycle management, including versioning More ❯
implementation and maintenance. Strong proficiency in SQL, experience with Python, exposure to DBT considered a plus Experience with AWS cloud computing services (Redshift, S3), GCP or similar Experience with Apache Airflow or similar nice to have Ability to merge the multiple requirements of data projects into robust future-proof solutions. Excellent written and verbal communications skills, also with non More ❯
in Python with libraries like TensorFlow, PyTorch, or Scikit-learn for ML, and Pandas, PySpark, or similar for data processing. Experience designing and orchestrating data pipelines with tools like Apache Airflow, Spark, or Kafka. Strong understanding of SQL, NoSQL, and data modeling. Familiarity with cloud platforms (AWS, Azure, GCP) for deploying ML and data solutions. Knowledge of MLOps practices More ❯
in Python with libraries like TensorFlow, PyTorch, or Scikit-learn for ML, and Pandas, PySpark, or similar for data processing. Experience designing and orchestrating data pipelines with tools like Apache Airflow, Spark, or Kafka. Strong understanding of SQL, NoSQL, and data modeling. Familiarity with cloud platforms (AWS, Azure, GCP) for deploying ML and data solutions. Knowledge of MLOps practices More ❯
and Workflows, Cluster Management, Catalog Design and Maintenance, Apps, Hive Metastore Management, Network Management, Delta Sharing, Dashboards, and Alerts. Proven experience working with big data technologies, i.e., Databricks and Apache Spark. Proven experience working with Azure data platform services, including Storage, ADLS Gen2, Azure Functions, Kubernetes. Background in cloud platforms and data architectures, such as Corporate DataLake, Medallion Architecture More ❯
in Python with libraries like TensorFlow, PyTorch, or Scikit-learn for ML, and Pandas, PySpark, or similar for data processing. Experience designing and orchestrating data pipelines with tools like Apache Airflow, Spark, or Kafka. Strong understanding of SQL, NoSQL, and data modeling. Familiarity with cloud platforms (AWS, Azure, GCP) for deploying ML and data solutions. Knowledge of MLOps practices More ❯
London, England, United Kingdom Hybrid / WFH Options
Scope3
and GraphQL APIs React w/Next.js for frontend applications Low latency + high throughput Golang API Big Query Data warehouse Airflow for batch orchestration Temporal for event orchestration Apache Beam (dataflow runner) for some batch jobs Most transformations are performed via SQL directly in Big Query. The Role We are excited to add a Lead Engineer to our More ❯
london (city of london), south east england, united kingdom
Tadaweb
in Python with libraries like TensorFlow, PyTorch, or Scikit-learn for ML, and Pandas, PySpark, or similar for data processing. Experience designing and orchestrating data pipelines with tools like Apache Airflow, Spark, or Kafka. Strong understanding of SQL, NoSQL, and data modeling. Familiarity with cloud platforms (AWS, Azure, GCP) for deploying ML and data solutions. Knowledge of MLOps practices More ❯
and Workflows, Cluster Management, Catalog Design and Maintenance, Apps, Hive Metastore Management, Network Management, Delta Sharing, Dashboards, and Alerts. Proven experience working with big data technologies, i.e., Databricks and Apache Spark. Proven experience working with Azure data platform services, including Storage, ADLS Gen2, Azure Functions, Kubernetes. Background in cloud platforms and data architectures, such as Corporate DataLake, Medallion Architecture More ❯
full-stack technologies including; Java, JavaScript, TypeScript, React, APIs, MongoDB, Elastic Search, DMN, BPMN and Kubernetes leverage data-streaming technologies including Kafka CDC, Kafka topic and related technologies, EMS, Apache Flink be able to innovate and incubate new ideas, have an opportunity to work on a broad range of problems, often dealing with large data sets, including real-time More ❯