/OpenShift including high availability Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
scale data processing and analytics Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
Ensure systems meet business requirements and industry practices for data integrity and quality. Manage ETL and ELT pipelines across many data sources (CSV/parquet files, API endpoints, etc) Design and build data models for the business end users. Write complex SQL queries for standard as well as ad more »
guildford, south east england, United Kingdom Hybrid / WFH Options
Hawksworth
Engineer or in a similar role Experience with data modelling, data warehousing and ETL frameworks Proficiency in working with relational databases (e.g., Oracle, PostgreSQL), Parquet/Delta files and big data technologies (e.g. Synapse, Hadoop, Spark, Kafka) Experience working with Microsoft Azure and associated data services Strong analytical and more »
experience working as a Software Engineer ideally using Rust Outstanding CompSci Fundamentals needed! Knowledge of data engineering principles/data lake architectures - Arrow/Parquet etc. Excellent knowledge of low-level concepts (e.g. compilers, x86, LLVM etc. more »
experience working as a Software Engineer ideally using Rust Outstanding CompSci Fundamentals needed! Knowledge of data engineering principles/data lake architectures - Arrow/Parquet etc. Excellent knowledge of low-level concepts (e.g. compilers, x86, LLVM etc. more »
is redeveloping our teams data platform from a legacy system to one based on AWS EMR, with Amazon RDS and DynamoDB ingestion converted to Parquet files, interrogatable through Spark and MapReduce. This modern platform will support rapid data insight generation, data experiments for new product development, our live Machine … Amazon DynamoDB, AWS Glue, MapReduce, Hive, Spark, YARN, Airflow. Ability to work with a range of structured, semi-structured and unstructured file formats including Parquet, json, csv, pdf, jpg. Accomplished data wrangling and scripting abilities, with advanced knowledge of Python, Linux shell scripting, SQL and NoSQL. Tools and methods more »
ROLE SUMMARY DATA ENGINEER REMOTE/TRAVEL TO CLIENT SITE IN LONDON Valcon UK are looking to recruit a self-motivated, highly logical and intellectually curious Data Engineer’s to join our expanding Data capability in the UK. As a more »
ROLE SUMMARY DATA ENGINEER REMOTE/TRAVEL TO CLIENT SITE IN LONDON Valcon UK are looking to recruit a self-motivated, highly logical and intellectually curious Data Engineer’s to join our expanding Data capability in the UK. As a more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC [Required] Experience in extracting and developing technical requirements from business goals and needs. [Required] Experience in solution integration and operability. [Required] Experience working more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC [Required] Experience in extracting and developing technical requirements from business goals and needs. [Required] Experience in solution integration and operability. [Required] Experience working more »
new technologies and frameworks Nice to have: Knowledge of databases, SQL Familiarity with Boost ASIO Familiarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, Flatbuffers Experience with gRPC, http/REST and Websocket protocols Experience with Google Cloud/AWS and/or containerization in more »
define objectives of data science projects, supported by senior colleagues if necessary. Extract, process and transform data in varied formats and platforms (e.g. csv, parquet, database), adopting appropriate tools, techniques, and applying best practice. Deliver, with minimal supervision, high quality data science outputs and products, for example reports, dashboards more »
relational models, and formats to support geo-enabled applications and graph database structures, using knowledge of various data formats including XML, JSON, flat file, Parquet etc. You can also manage metadata across your development teams systems and beyond, with a practitioner knowledge of tracking data provenance and data lineage. more »
london, south east england, United Kingdom Hybrid / WFH Options
Careers at MI5, SIS and GCHQ
relational models, and formats to support geo-enabled applications and graph database structures, using knowledge of various data formats including XML, JSON, flat file, Parquet etc. You can also manage metadata across your development teams systems and beyond, with a practitioner knowledge of tracking data provenance and data lineage. more »
Chicago, Illinois, United States Hybrid / WFH Options
Request Technology - Robyn Honquest
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Education more »
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience more »
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience more »
flows & pipelines in a cloud environment using a progressive tech stack e.g. Databricks, Spark, Python, PySpark, Delta Lake, SQL, Logic Apps, Azure Functions, ADLS, Parquet, Neo4J, Flask Ingest and integrate data from a large number of disparate data sources Design and build complex data models for analytical insight Write more »
flows & pipelines in a cloud environment using a progressive tech stack e.g. Databricks, Spark, Python, PySpark, Delta Lake, SQL, Logic Apps, Azure Functions, ADLS, Parquet, Neo4J, Flask Ingest and integrate data from a large number of disparate data sources Design and build complex data models for analytical insight Write more »
unstructured data, extracting information and identifying linkages across disparate datasets. • Expert in creating data structures optimized for storage and various query patterns for e.g. Parquet and Delta Lake • Experience in traditional data warehousing concepts (Kimball Methodology, Star Schema, SCD)/ETL tools (Azure Data factory, Informatica) • Experience in data more »
london, south east england, United Kingdom Hybrid / WFH Options
trg.recruitment
/OpenShift including high availability Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
london, south east england, United Kingdom Hybrid / WFH Options
trg.recruitment
scale data processing and analytics Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
new technologies and frameworks Nice to have: Knowledge of databases, SQL Familiarity with Boost ASIO Familiarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, Flatbuffers Experience with gRPC, http/REST and Websocket protocols Experience with Google Cloud/AWS and/or containerization in more »
of data modelling (particularly star-schema) and can produce, maintain, and update relevant data models for specific business needs including bespoke serialization techniques (e.g. Parquet) and table formats (Delta etc) This is an incredibly exciting role that rarely becomes available. You will own a well-designed data lake serving more »