and previous experience in early stage startups. Experience with CI and CD tools such as GitHub actions. Understanding of Big Data file formats (e.g. Parquet, ORC, Avro). Understanding of Big Data processing engine internals. Since we are a remote company, we are open to candidates from overseas. No more »
/OpenShift including high availability Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
scale data processing and analytics Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
up and learn new technologies and frameworksNice to have:Knowledge of databases, SQLFamiliarity with Boost ASIOFamiliarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, FlatbuffersExperience with gRPC, http/REST and Websocket protocolsExperience with Google Cloud/AWS and/or containerization in Docker/ more »
unstructured data, extracting information and identifying linkages across disparate datasets. • Expert in creating data structures optimized for storage and various query patterns for e.g. Parquet and Delta Lake • Experience in traditional data warehousing concepts (Kimball Methodology, Star Schema, SCD)/ETL tools (Azure Data factory, Informatica) • Experience in data more »
Ensure systems meet business requirements and industry practices for data integrity and quality. Manage ETL and ELT pipelines across many data sources (CSV/parquet files, API endpoints, etc) Design and build data models for the business end users. Write complex SQL queries for standard as well as ad more »
new technologies and frameworks Nice to have: Knowledge of databases, SQL Familiarity with Boost ASIO Familiarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, Flatbuffers Experience with gRPC, http/REST and Websocket protocols Experience with Google Cloud/AWS and/or containerization in more »
experience working as a Software Engineer ideally using Rust Outstanding CompSci Fundamentals needed! Knowledge of data engineering principles/data lake architectures - Arrow/Parquet etc. Excellent knowledge of low-level concepts (e.g. compilers, x86, LLVM etc. more »
define objectives of data science projects, supported by senior colleagues if necessary. Extract, process and transform data in varied formats and platforms (e.g. csv, parquet, database), adopting appropriate tools, techniques, and applying best practice. Deliver, with minimal supervision, high quality data science outputs and products, for example reports, dashboards more »
service ad-hoc analysis. * Develop and manage data pipelines using Azure Synapse Analytics or Azure Data Factory. * Work with columnar storage formats such as Parquet and Delta to optimize data storage and retrieval processes. * Design, develop, and maintain Power BI reports and dashboards to meet business needs. * Implement DevOps … very important!), with the ability to effectively collaborate with cross-functional teams and customers Other skills * Experience or knowledge of columnar storage formats, especially Parquet and Delta * Familiarity with DevOps practices, particularly source control using Git. * Strong analytical and problem-solving abilities. * Ability to prioritize and manage multiple tasks more »
Employment Type: Permanent
Salary: £65000 - £70000/annum Hybrid, Health, Dental, Extra Hols