A-Level grades are included on your CV). Basic scripting knowledge in Python or Bash Excellent customer-facing skills You have a sales spark - while this role isn't a focussed sales role, this is required due to the nature of the role A motivated self-starter with More ❯
Connected Plant, Emerson Plantweb/AMS, GE/Meridum APM, Aveva, Bentley, and OSIsoft PI Familiarity with relevant technology, such as Big Data (Hadoop, Spark, Hive, BigQuery); Data Warehouses; Business Intelligence; and Machine Learning Savvy at helping customers create business cases with quantified ROI to justify new investments Experience More ❯
and ensure efficient integration into Feast feature store. Requirements Good knowledge of programming languages such as Python or Java. Strong experience with streaming technologies (Spark, PySpark, Flink, KSQL or similar) for developing data transformation pipelines. Solid understanding and practical experience with SQL and relational databases (PostgreSQL preferred). Proficiency … with AWS EMR for running and managing Spark workloads. Experience in SQL-based transformation workflows, particularly using DBT in BigQuery. Experience with Terraform to define and manage cloud infrastructure through code. Desirables: Familiarity with AWS, GCP or other cloud providers. Experience with containerisation technologies (Docker, Kubernetes). Familiarity with More ❯
a Data Engineer II, you should be an expert with data lake fundamentals around storage, compute, etc., familiar with multiple data processing stacks like Spark, and core data concepts (e.g. Data Modeling). You should have deep understanding of the architecture for enterprise level data lake/mesh solutions … systems as it pertains to data storage and computing Experience with Redshift, Oracle, NoSQL etc. Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Bachelor's degree PREFERRED QUALIFICATIONS Experience working on and delivering end to end projects independently Experience providing technical leadership and mentoring other engineers More ❯
Role type: Perm Location: Hybrid/Remote from within the UK or Greece Preferred start date: ASAP LIFE AT SATALIA As an organisation, we push the boundaries of data science, optimisation and artificial intelligence to solve the hardest problems in More ❯
world architectures that business engineering teams buy into and build their applications around. Required Qualifications, Capabilities, and Skills: Experience across the data lifecycle with Spark-based frameworks for end-to-end ETL, ELT & reporting solutions using key components like Spark SQL & Spark Streaming. Strong knowledge of multi … fantastic end-to-end engineering experience supported by excellent tooling and automation. Preferred Qualifications, Capabilities, and Skills: Good understanding of the Big Data stack (Spark/Iceberg). Ability to learn new technologies and patterns on the job and apply them effectively. Good understanding of established patterns, such as More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook. Benefits At Databricks, we strive to provide comprehensive More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. To learn more, follow Databricks on Twitter ,LinkedIn and Facebook . Benefits At Databricks, we strive to provide More ❯
science/quantitative modeling to real world, financial use cases. Knowledge of open-source technologies and platforms commonly used for data analysis (e.g., Hadoop, Spark, etc.). More ❯
conferences or journals - 3+ years of hands-on predictive modeling and large data analysis experience - Experience working with large-scale distributed systems such as Spark, Sagemaker or similar frameworks Amazon is committed to a diverse and inclusive workplace. Amazon is an equal opportunity employer and does not discriminate on More ❯
production issues. What We Value Experience with monitoring systems using tools like Prometheus and writing health checks Interest in learning and managing technologies like Spark, Hadoop, Elasticsearch, and Cassandra Familiarity with deploying GPUs Moderate experience with TCP/IP networking Ability to work independently with minimal supervision Ability to More ❯
Guildford, England, United Kingdom Hybrid / WFH Options
Crowd Connected
any environment. On average every day 10,000 new app downloads from the app stores contain our code. Our innovative solutions create meaningful connections, spark real change, and drive success for businesses around the world. If you’re ready to be part of a team shaping the future of More ❯
building and managing real-time data pipelines across a track record of multiple initiatives. Expertise in developing data backbones using distributed streaming platforms (Kafka, Spark Streaming, Flink, etc.). Experience working with cloud platforms such as AWS, GCP, or Azure for real-time data ingestion and storage. Ability to … or a similar language. Proficiency in database technologies (SQL, NoSQL, time-series databases) and data modelling. Strong understanding of data pipeline orchestration tools (e.g., Apache Airflow, Kubernetes). You thrive when working as part of a team Comfortable in a fast-paced environment Have excellent written and verbal English More ❯
technologies such as SQL data warehouses, while at other times you will be helping them to discover and implement the most cutting edge tools; Spark on Kubernetes, cloud-based elastic compute engines, and GPUs. If you are interested in staying at the bleeding edge of big data and AI … customer requirements in terms of scalability, availability and security and provide architecture recommendations. Deploy Dataiku in a large variety of technical environments (SaaS, Kubernetes, Spark, Cloud or on-prem). Automate operation, installation, and monitoring of the Data Science ecosystem components in our infrastructure stack. Collaborate with Revenue and … pre and post sale. What will make you stand out: Some knowledge in Data Science and/or machine learning. Hands-on experience with Spark ecosystem for setup, administration, troubleshooting and tuning. Experience with authentication and authorization systems like (A)AD, IAM, and LDAP. What does the hiring process More ❯
economic and data science models into production. We are looking for a individual who is interested in working with the latest big data technology (Spark, EMR, Glue, SageMaker, and Airflow) and collaborate with Economist and Scientist in creating scalable solutions for our multiple Retail Businesses. Key job responsibilities - Partnering … as Python, Java, Scala, or NodeJS - Experience mentoring team members on best practices PREFERRED QUALIFICATIONS - Experience with big data technologies such as: Hadoop, Hive, Spark, EMR Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation More ❯
ECOM are pleased to be exclusively recruiting for a Senior Data Engineer here in Manchester. You'll join a team where your work reaches millions. This role is within a forward-thinking company leading, offering a dynamic environment where you More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, ApacheSpark, Delta Lake and MLflow. Benefits At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of More ❯
problem. Right now, we use: A variety of languages, including Java and Go for backend and Typescript for frontend Open-source technologies like Cassandra, Spark, Elasticsearch, React, and Redux Industry-standard build tooling, including Gradle, Webpack, and GitHub What We Value Ability to communicate and collaborate with a variety More ❯
Computer Science, Engineering, Mathematics, or a related field - Data Warehousing experience with Redshift, Teradata. - Experience with workflow management platforms for data engineering pipelines (ex. Apache Airflow) - Experience with Big Data Technologies (Spark, Hadoop, Hive, Pig, etc.) - Experience building/operating highly available, distributed systems of data extraction, ingestion More ❯
quality data solutions. Automation: Implement automation processes and best practices to streamline data workflows and reduce manual interventions. Must have: AWS, ETL, EMR, GLUE, Spark/Scala, Java, Python. Good to have: Cloudera - Spark, Hive, Impala, HDFS, Informatica PowerCenter, Informatica DQ/DG, Snowflake Erwin. Qualifications: Bachelor's More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯
efficient data models for real-time analytics. Proven experience in managing real-time data pipelines across multiple initiatives. Expertise in distributed streaming platforms (Kafka, Spark Streaming, Flink). Experience with GCP (preferred), AWS, or Azure for real-time data ingestion and storage. Strong programming skills in Python, Java, or … Scala . Proficiency in SQL, NoSQL, and time-series databases . Knowledge of orchestration tools (Apache Airflow, Kubernetes). If you are a passionate and experienced Senior Data Engineer seeking a Lead role, or a Lead Data Engineer aiming to make an impact like in a previous position, we More ❯