primarily GCP. Experience with some or all of the services below would put you at the top of our list Google Cloud Storage Google Data Transfer Service Google Dataflow (ApacheBeam) Google PubSub Google CloudRun BigQuery or any RDBMS Python Debezium/Kafka dbt (Data Build tool) Interview process Interviewing is a two way process and we want More ❯
basic security principles at infrastructure and application level. You have knowledge of cloud based ML solutions from GCP or AWS Experience with streaming data processing frameworks such as Flink, Beam, Spark, Kafka Streams Experience with Ansible, Terraform, GitHub Actions, Infrastructure as Code, AWS or other cloud ecosystems Knowledge/interest in payment platforms, foreign exchange & complex systems architecture Be More ❯
learning systems at scale You have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, ApacheBeam, or Spark. You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative teams Experience with TensorFlow … and/or Google Cloud Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like ApacheBeam/Spark is a plus Where You'll Be For this role you should be based in London (UK). #J-18808-Ljbffr More ❯
best practices Stay current with the evolving data landscape, identifying opportunities for improvement across the data platform Core Tech Stack & Skills Experience with GCP services including BigQuery, Dataflow (or ApacheBeam), Pub/Sub, and Cloud Composer (or Apache Airflow) Strong command of Python and experience working with dbt Understanding of stream processing and designing for both More ❯
and join a high-energy, positive team environment! What You'll Do Work with large-scale data pipelines with data processing frameworks like Scio, BigQuery, Google Cloud Platform and Apache Beam. Develop, deploy, and operate Java services that impact millions of users. Work towards supporting machine learning projects powering the experience that suits each user individually. Collaborate with other … constructive accountability. Who You Are You are familiar with the concepts of data modeling, data access, and data storage techniques. You are familiar with distributed data processing frameworks (ex: Beam, Spark). You want to work on a team employing agile software development processes, data-driven development, and responsible experimentation. You value opportunities to work collaboratively. Where You'll More ❯
London, England, United Kingdom Hybrid / WFH Options
Scope3
and GraphQL APIs React w/Next.js for frontend applications Low latency + high throughput Golang API Big Query Data warehouse Airflow for batch orchestration Temporal for event orchestration ApacheBeam (dataflow runner) for some batch jobs Most transformations are performed via SQL directly in Big Query. The Role We are excited to add a Lead Engineer to More ❯
London, England, United Kingdom Hybrid / WFH Options
Spotify
constructive feedback to foster accountability, growth, and collaboration within the team. Who You Are Experienced with Data Processing Frameworks: Skilled with higher-level JVM-based frameworks such as Flink, Beam, Dataflow, or Spark. Comfortable with Ambiguity: Able to work through loosely defined problems and thrive in autonomous team environments. Skilled in Cloud-based Environments: Proficient with large-scale data More ❯
London, England, United Kingdom Hybrid / WFH Options
Spotify AB
constructive feedback to foster accountability, growth, and collaboration within the team. Who You Are Experienced with Data Processing Frameworks: Skilled with higher-level JVM-based frameworks such as Flink, Beam, Dataflow, or Spark. Comfortable with Ambiguity: Able to work through loosely defined problems and thrive in autonomous team environments. Skilled in Cloud-based Environments: Proficient with large-scale data More ❯
its nucleus. Oversimplified, we encode, process, and model biometric and survey data. Managing and optimizing this process E2E is your remit. We’re currently migrating our pipelines to use Beam/DataFlow with a BigQuery sink and shifting our DB from Postgres to BigQuery. From there, we have lots of value to extract from our existing data, lots of More ❯
for large datasets. Expertise in BigQuery, including advanced SQL, partitioning, clustering, and performance tuning. Hands-on experience with at least one of the following GCP data processing services: Dataflow (ApacheBeam), Dataproc (Apache Spark/Hadoop), or Composer (Apache Airflow). Proficiency in at least one scripting/programming language (e.g., Python, Java, Scala) for data More ❯
Location Remote, with occasional company meetings in Bristol (maximum 1x a month) Beam Connectivity are a startup in the automotive IoT space. We work with established and up-and-coming vehicle manufacturers to deliver best-in-class connected vehicle experiences. After a successful first 5 years and announcement of strategic funding from Fortescue, we have now secured new projects … which have a deep understanding, nor the responsibility for the end-to-end system. Delivering a robust automotive IoT solution requires a wide variety of skills and experience. At Beam, we are a truly multi-disciplinary team, covering all the skills required to deliver a first class connected experience. Our flagship product is the Connected Vehicle as a Service … We do all this under one roof, so you’ll be exposed to all this technology at one time or other. This should excite you, not scare you... At Beam, we spend our engineering energy on three main things: Building out our core CVaaS platform - Building new features, adding resilience, and rolling this out to vehicle manufacturers. Research projects More ❯
learning systems at scale You have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, ApacheBeam, or Spark You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative teams Experience with TensorFlow … pyTorch, and/or other scalable Machine learning frameworks Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like ApacheBeam/Spark Where You'll Be We offer you the flexibility to work where you work best! For this role, you can be within the EMEA region More ❯
Lexington, Massachusetts, United States Hybrid / WFH Options
Equiliem
Qualifications: • Bachelor's Degree in Computer Science. Recent graduates or candidates without a Bachelor's degree considered with clear evidence of significant outside-of-classroom experience. • Experience with the Apache Maven or Gradle build system. • Ability to understand front-end source code written in React or similar frameworks. Provide guidance to less experienced front-end engineers. • General knowledge of … and reinforcement learning concepts, frameworks, and environments, such as Pandas, TensorFlow, and Jupyter Notebook. • Broad knowledge of the general features, capabilities, and trade-offs of common data warehouse (e.g. Apache Hadoop); workflow orchestration (e.g. ApacheBeam); data extract, transform and load (ETL); and stream processing (e.g. Kafka) technologies. Hands-on experience with several of these technologies. -This More ❯
learning systems at scale You have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, ApacheBeam, or Spark You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative teams Experience with TensorFlow … pyTorch, and/or other scalable Machine learning frameworks Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like ApacheBeam/Spark Where You'll Be We offer you the flexibility to work where you work best! For this role, you can be within the EMEA region More ❯
TensorFlow, PyTorch, scikit-learn) Experience with cloud platforms (AWS, GCP, Azure) Experience with CI/CD pipelines for machine learning (e.g., Vertex AI) Familiarity with data processing tools like ApacheBeam/Dataflow Strong understanding of monitoring and maintaining models in production environments Experience with containerization tools (e.g., Docker) Problem-solving skills with the ability to troubleshoot model More ❯
data for ML. Who You Are You have proven experience in data engineering, including creating reliable, efficient, and scalable data pipelines using data processing frameworks such as Scio, DataFlow, Beam or equivalent. You are comfortable working with large datasets using SQL and data analytics platforms such as BigQuery. You are knowledgeable in cloud-based environments, preferably with an understanding More ❯
Python, or similar languages. Experience with TensorFlow, PyTorch, Scikit-learn, etc. is a strong plus. You have some experience with large scale, distributed data processing frameworks/tools like ApacheBeam, Apache Spark, or even our open source API for it - Scio, and cloud platforms like GCP or AWS. You care about agile software processes, data-driven More ❯
end tech specs and modular architectures for ML frameworks in complex problem spaces in collaboration with product teams Experience with large scale, distributed data processing frameworks/tools like ApacheBeam, Apache Spark, and cloud platforms like GCP or AWS Where You'll Be We offer you the flexibility to work where you work best! For this More ❯
and CNNs Excellent communication skills Degree in CS, maths, statistics, engineering, physics or similar Desirable Requirements: NoSQL databases - Elasticsearch, MongoDB etc (bonus) Modern Data tools such as Spark/Beam (bonus) Streaming technologies such as Spark/Akka Streams (bonus) Tagged as: Industry , NLP , United Kingdom More ❯
production-oriented engineering mindset Strong software engineering skills with experience in building distributed systems Expertise in Python and experience with distributed computing frameworks Experience building distributed ETL pipelines in ApacheBeam, Spark, or Ray Nice to have Have built a petabyte-scale data-engine to handle a fleet of robots. Experience with the following technologies: Rust, Go, Vector More ❯