technologies such as SQL data warehouses, while at other times you will be helping them to discover and implement the most cutting edge tools; Spark on Kubernetes, cloud-based elastic compute engines, and GPUs. If you are interested in staying at the bleeding edge of big data and AI … customer requirements in terms of scalability, availability and security and provide architecture recommendations. Deploy Dataiku in a large variety of technical environments (SaaS, Kubernetes, Spark, Cloud or on-prem). Automate operation, installation, and monitoring of the Data Science ecosystem components in our infrastructure stack. Collaborate with Revenue and … pre and post sale. What will make you stand out: Some knowledge in Data Science and/or machine learning. Hands-on experience with Spark ecosystem for setup, administration, troubleshooting and tuning. Experience with authentication and authorization systems like (A)AD, IAM, and LDAP. What does the hiring process More ❯
economic and data science models into production. We are looking for a individual who is interested in working with the latest big data technology (Spark, EMR, Glue, SageMaker, and Airflow) and collaborate with Economist and Scientist in creating scalable solutions for our multiple Retail Businesses. Key job responsibilities - Partnering … as Python, Java, Scala, or NodeJS - Experience mentoring team members on best practices PREFERRED QUALIFICATIONS - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation More ❯
ECOM are pleased to be exclusively recruiting for a Senior Data Engineer here in Manchester. You'll join a team where your work reaches millions. This role is within a forward-thinking company leading, offering a dynamic environment where you More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
of Data Science Lead the data science strategy and team to deliver data science solutions e.g. retention, acquisitions and customer management using Python and Spark Lead the hiring to build a great pool of Data Scientists and Engineers for the team and support the recruitment activities of other data More ❯
problem. Right now, we use: A variety of languages, including Java and Go for backend and Typescript for frontend Open-source technologies like Cassandra, Spark, Elasticsearch, React, and Redux Industry-standard build tooling, including Gradle, Webpack, and GitHub What We Value Ability to communicate and collaborate with a variety More ❯
Computer Science, Engineering, Mathematics, or a related field - Data Warehousing experience with Redshift, Teradata. - Experience with workflow management platforms for data engineering pipelines (ex. Apache Airflow) - Experience with Big Data Technologies (Spark, Hadoop, Hive, Pig, etc.) - Experience building/operating highly available, distributed systems of data extraction, ingestion More ❯
quality data solutions. Automation: Implement automation processes and best practices to streamline data workflows and reduce manual interventions. Must have: AWS, ETL, EMR, GLUE, Spark/Scala, Java, Python. Good to have: Cloudera - Spark, Hive, Impala, HDFS, Informatica PowerCenter, Informatica DQ/DG, Snowflake Erwin. Qualifications: Bachelor's More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯
in Scala, Python and/or Java. Strong experience with SQL, including querying, optimizing, and managing databases. Experience with data processing platforms such as Spark, Hadoop. Demonstrated experience with GCP services such as DataProc, BigQuery, GCS, IAM, and others, and/or their AWS equivalents. Work well as an … implement elegant solutions for them. Are a data enthusiast who wants to be surrounded by brilliant teammates and huge challenges. Bonus Points: Experience with Apache Airflow, including designing, managing, and troubleshooting DAGs and data pipelines. Experience with CI/CD pipelines and tools like Jenkins, including automating the process More ❯
Vision. Responsibilities: Design, develop, and productionize machine learning models across various applications. Work with Python (ideally production-level code) and other tools like SQL, Spark, and Databricks. Apply clustering, classification, regression, time series modelling, NLP, and deep learning. Develop recommendation engines and leverage third-party data enhancements. Implement MLOps … data-driven solutions. Requirements: MSc or PhD Degree in Computer Science, Artificial Intelligence, Mathematics, Statistics or related fields. Strong Python skills (bonus: C++, SQL, Spark) Experience in ML algorithms (XGBoost, clustering, regression) Expertise in Time Series, NLP, Computer Vision, MLOps Knowledge of AWS/Azure/GCP, CI/ More ❯
strongly preferred; other languages include Java, Scala, TypeScript, C++, C#). Experience using big data technologies in cloud environments to build data pipelines (e.g. Spark, EMR, Lambda, etc.). Excellent communication, organization, and prioritization skills, with a strong ability to deliver results within tight timelines. Passionate about working with … to ensure secure and efficient data operations that support business growth and strategic objectives. Writing code - lots of it! We use Python, Java, TypeScript, Spark, and SQL, welcoming engineers from diverse programming backgrounds who are passionate about building robust data solutions. Design, architect, and implement scalable, maintainable data pipelines More ❯
to ensure secure and efficient data operations that support business growth and strategic objectives. Writing code - lots of it! We use Python, Java, TypeScript, Spark, and SQL, welcoming engineers from diverse programming backgrounds who are passionate about building robust data solutions. Design, architect, and implement scalable, maintainable data pipelines … preferred with other languages including Java, Scala, TypeScript, C++, C#). Experience using big data technologies in cloud environments to build data pipelines (e.g. Spark, EMR, Lambda etc.). Excellent communication, organisation and prioritisation skills, and have a strong ability to deliver results within tight timelines. Passionate to work More ❯
Do you ever have the urge to do things better than the last time? We do. And it's this urge that drives us every day. Our environment of discovery and innovation means we're able to create deep and More ❯
have experience architecting data pipelines and are self-sufficient in getting the data you need to build and evaluate models, using tools like Dataflow, Apache Beam, or Spark. You care about agile software processes, data-driven development, reliability, and disciplined experimentation You have experience and passion for fostering collaborative … Platform is a plus Experience with building data pipelines and getting the data you need to build and evaluate your models, using tools like Apache Beam/Spark is a plus Where You'll Be This role is based in London (UK). We offer you the flexibility More ❯
Informatics, Information Systems, or another quantitative field. They should also have experience using the following software/tools: Experience with big data tools: Hadoop, Spark, Kafka, etc. Experience with relational SQL and NoSQL databases, including Postgres and Cassandra. Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow … etc. Experience with AWS cloud services: EC2, EMR, RDS, Redshift. Experience with stream-processing systems: Storm, Spark-Streaming, etc. Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc. Salary: 30000 per annum + benefits Apply For This Job If you would like to apply More ❯
With us, you'll do meaningful work from Day 1. Our collaborative culture is built on three core behaviors: We Play to Win, We Get Better Every Day & We Succeed Together. And we mean it - we want you to grow More ❯
As a Senior BI Developer, you will be at the forefront of creating Analytical Solutions and insights into a wide range of business processes throughout the organisation and playing a core role in our strategic initiatives to enhance data-driven More ❯
About Us The Company: Dotdigital is a thriving global community of passionate, dedicated professionals, committed to the collective success of the organization and its clients. Our core principles of innovation, teamwork, and client-focused solutions drive us to approach challenges More ❯
areas of Data Mining, Classical Machine Learning, Deep Learning, NLP and Computer Vision. Experience with Large Scale/Big Data technology, such as Hadoop, Spark, Hive, Impala, PrestoDb. Hands-on capability developing ML models using open-source frameworks in Python and R and applying them on real client use … cases. Proficient in one of the deep learning stacks such as PyTorch or Tensorflow. Working knowledge of parallelisation and async paradigms in Python, Spark, Dask, Apache Ray. An awareness and interest in economic, financial and general business concepts and terminology. Excellent written and verbal command of English. Strong More ❯
Join our dynamic team at Baseten, where we're revolutionizing AI deployment with cutting-edge inference infrastructure. Backed by premier investors such as IVP , Spark Capital , Greylock , and Conviction , we're trusted by leading enterprises and AI-driven innovators-including Descript , Bland.ai , Patreon , Writer , and Robust Intelligence -to deliver More ❯