and ensure efficient integration into Feast feature store. Requirements Good knowledge of programming languages such as Python or Java. Strong experience with streaming technologies (Spark, PySpark, Flink, KSQL or similar) for developing data transformation pipelines. Solid understanding and practical experience with SQL and relational databases (PostgreSQL preferred). Proficiency … with AWS EMR for running and managing Spark workloads. Experience in SQL-based transformation workflows, particularly using DBT. Experience with Terraform to define and manage cloud infrastructure through code. Desirables: Familiarity with GCP. Experience with containerisation technologies (Docker, Kubernetes). Familiarity with streaming data ingestion technologies (Kafka, Debezium). More ❯
Role type: Perm Location: Hybrid/Remote from within the UK or Greece Preferred start date: ASAP LIFE AT SATALIA As an organisation, we push the boundaries of data science, optimisation and artificial intelligence to solve the hardest problems in More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
of Data Science Lead the data science strategy and team to deliver data science solutions e.g. retention, acquisitions and customer management using Python and Spark Lead the hiring to build a great pool of Data Scientists and Engineers for the team and support the recruitment activities of other data More ❯
of Data Science Lead the data science strategy and team to deliver data science solutions e.g. retention, acquisitions and customer management using Python and Spark Lead the hiring to build a great pool of Data Scientists and Engineers for the team and support the recruitment activities of other data More ❯
problem. Right now, we use: A variety of languages, including Java and Go for backend and Typescript for frontend Open-source technologies like Cassandra, Spark, Elasticsearch, React, and Redux Industry-standard build tooling, including Gradle, Webpack, and GitHub What We Value Ability to communicate and collaborate with a variety More ❯
building and managing real-time data pipelines across a track record of multiple initiatives. Expertise in developing data backbones using distributed streaming platforms (Kafka, Spark Streaming, Flink, etc.). Experience working with cloud platforms such as AWS, GCP, or Azure for real-time data ingestion and storage. Ability to … or a similar language. Proficiency in database technologies (SQL, NoSQL, time-series databases) and data modelling. Strong understanding of data pipeline orchestration tools (e.g., Apache Airflow, Kubernetes). You thrive when working as part of a team Comfortable in a fast-paced environment Have excellent written and verbal English More ❯
technologies such as SQL data warehouses, while at other times you will be helping them to discover and implement the most cutting edge tools; Spark on Kubernetes, cloud-based elastic compute engines, and GPUs. If you are interested in staying at the bleeding edge of big data and AI … customer requirements in terms of scalability, availability and security and provide architecture recommendations. Deploy Dataiku in a large variety of technical environments (SaaS, Kubernetes, Spark, Cloud or on-prem). Automate operation, installation, and monitoring of the Data Science ecosystem components in our infrastructure stack. Collaborate with Revenue and … pre and post sale. What will make you stand out: Some knowledge in Data Science and/or machine learning. Hands-on experience with Spark ecosystem for setup, administration, troubleshooting and tuning. Experience with authentication and authorization systems like (A)AD, IAM, and LDAP. What does the hiring process More ❯
strongly preferred; other languages include Java, Scala, TypeScript, C++, C#). Experience using big data technologies in cloud environments to build data pipelines (e.g. Spark, EMR, Lambda, etc.). Excellent communication, organization, and prioritization skills, with a strong ability to deliver results within tight timelines. Passionate about working with … to ensure secure and efficient data operations that support business growth and strategic objectives. Writing code - lots of it! We use Python, Java, TypeScript, Spark, and SQL, welcoming engineers from diverse programming backgrounds who are passionate about building robust data solutions. Design, architect, and implement scalable, maintainable data pipelines More ❯
to ensure secure and efficient data operations that support business growth and strategic objectives. Writing code - lots of it! We use Python, Java, TypeScript, Spark, and SQL, welcoming engineers from diverse programming backgrounds who are passionate about building robust data solutions. Design, architect, and implement scalable, maintainable data pipelines … preferred with other languages including Java, Scala, TypeScript, C++, C#). Experience using big data technologies in cloud environments to build data pipelines (e.g. Spark, EMR, Lambda etc.). Excellent communication, organisation and prioritisation skills, and have a strong ability to deliver results within tight timelines. Passionate to work More ❯
Job Title: Gen AI Architect Duration: Full time Job Description: We are seeking a highly skilled Generative AI Architect to join our dynamic team and drive the development of advanced AI systems that have capabilities to learn, reason and make More ❯
Job Title: Gen AI Architect Duration: Full time Job Description: We are seeking a highly skilled Generative AI Architect to join our dynamic team and drive the development of advanced AI systems that have capabilities to learn, reason and make More ❯
areas of Data Mining, Classical Machine Learning, Deep Learning, NLP and Computer Vision. Experience with Large Scale/Big Data technology, such as Hadoop, Spark, Hive, Impala, PrestoDb. Hands-on capability developing ML models using open-source frameworks in Python and R and applying them on real client use … cases. Proficient in one of the deep learning stacks such as PyTorch or Tensorflow. Working knowledge of parallelisation and async paradigms in Python, Spark, Dask, Apache Ray. An awareness and interest in economic, financial and general business concepts and terminology. Excellent written and verbal command of English. Strong More ❯
Central London, London, United Kingdom Hybrid / WFH Options
167 Solutions Ltd
develop scalable solutions that enhance data accessibility and efficiency across the organisation. Key Responsibilities Design, build, and maintain data pipelines using SQL, Python, and Spark . Develop and manage data warehouse and lakehouse solutions for analytics, reporting, and machine learning. Implement ETL/ELT processes using tools such as … Apache Airflow, AWS Glue, and Amazon Athena . Work with cloud-native technologies to support scalable, serverless architectures. Collaborate with data science teams to streamline feature engineering and model deployment. Ensure data governance, lineage, and compliance best practices. Mentor and support team members in data engineering best practices . … Skills & Experience Required 6+ years of experience in data engineering within large-scale digital environments. Strong programming skills in Python, SQL, and Spark (SparkSQL) . Expertise in Snowflake and modern data architectures. Experience designing and managing data pipelines, ETL, and ELT workflows . Knowledge of AWS services such as More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯
in Scala, Python and/or Java. Strong experience with SQL, including querying, optimizing, and managing databases. Experience with data processing platforms such as Spark, Hadoop. Demonstrated experience with GCP services such as DataProc, BigQuery, GCS, IAM, and others, and/or their AWS equivalents. Work well as an … implement elegant solutions for them. Are a data enthusiast who wants to be surrounded by brilliant teammates and huge challenges. Bonus Points: Experience with Apache Airflow, including designing, managing, and troubleshooting DAGs and data pipelines. Experience with CI/CD pipelines and tools like Jenkins, including automating the process More ❯
Vision. Responsibilities: Design, develop, and productionize machine learning models across various applications. Work with Python (ideally production-level code) and other tools like SQL, Spark, and Databricks. Apply clustering, classification, regression, time series modelling, NLP, and deep learning. Develop recommendation engines and leverage third-party data enhancements. Implement MLOps … data-driven solutions. Requirements: MSc or PhD Degree in Computer Science, Artificial Intelligence, Mathematics, Statistics or related fields. Strong Python skills (bonus: C++, SQL, Spark) Experience in ML algorithms (XGBoost, clustering, regression) Expertise in Time Series, NLP, Computer Vision, MLOps Knowledge of AWS/Azure/GCP, CI/ More ❯
Do you ever have the urge to do things better than the last time? We do. And it's this urge that drives us every day. Our environment of discovery and innovation means we're able to create deep and More ❯
have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark. You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative … Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark is a plus Where You'll Be This role is based in London (UK). We offer you the flexibility More ❯
Join our dynamic team at Baseten, where we're revolutionizing AI deployment with cutting-edge inference infrastructure. Backed by premier investors such as IVP , Spark Capital , Greylock , and Conviction , we're trusted by leading enterprises and AI-driven innovators-including Descript , Bland.ai , Patreon , Writer , and Robust Intelligence -to deliver More ❯
City of London, London, United Kingdom Hybrid / WFH Options
Cathcart Technology
Prior Senior Data Scientist with Machine Learning experience ** Strong understanding and experience with ML models and ML observability tools ** Strong Python and SQL experience ** Spark/Apache Airflow ** ML frame work experience (PyTorch/TensorFlow/Scikit-Learn) ** Experience with cloud platforms (preferably AWS) ** Experience with containerisation technologies More ❯