development experience in at least one of the following: Java, Python, PySpark or Scala.Experience working with a variety of data formats such as JSON, Parquet, XML etc.Experience with or developed understanding of the application of ETL/ELT tools.Experience with NoSQL type environments, Data Lakes, Lake-Houses (Cassandra, MongoDB more »
flows & pipelines in a cloud environment using a progressive tech stack e.g. Databricks, Spark, Python, PySpark, Delta Lake, SQL, Logic Apps, Azure Functions, ADLS, Parquet, Neo4J, Flask Ingest and integrate data from a large number of disparate data sources Design and build complex data models for analytical insight Write more »
relational models, and formats to support geo-enabled applications and graph database structures, using knowledge of various data formats including XML, JSON, flat file, Parquet etc. You can also manage metadata across your development teams systems and beyond, with a practitioner knowledge of tracking data provenance and data lineage. more »
Guildford, England, United Kingdom Hybrid / WFH Options
Hawksworth
Engineer or in a similar role Experience with data modelling, data warehousing and ETL frameworks Proficiency in working with relational databases (e.g., Oracle, PostgreSQL), Parquet/Delta files and big data technologies (e.g. Synapse, Hadoop, Spark, Kafka) Knowledge of Microsoft Azure and associated data services is a good to more »
Ensure systems meet business requirements and industry practices for data integrity and quality. Manage ETL and ELT pipelines across many data sources (CSV/parquet files, API endpoints, etc) Design and build data models for the business end users. Write complex SQL queries for standard as well as ad more »
/OpenShift including high availability Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience more »
Chicago, Illinois, United States Hybrid / WFH Options
Request Technology - Robyn Honquest
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Education more »
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience more »
scale data processing and analytics Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
Chicago, Illinois, United States Hybrid / WFH Options
Request Technology
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, Pytest, etc.). Bonus points Experience with Scripting languages such as more »
command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC Experience in solution integration and operability. Experience working with Infrastructure Technologies and Teams. Experience using Service-now or similar more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC Experience in solution integration and operability. Experience working with Infrastructure Technologies and Teams. Experience using Service-now or similar more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC [Required] Experience in extracting and developing technical requirements from business goals and needs. [Required] Experience in solution integration and operability. [Required] Experience working more »
tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC [Required] Experience in extracting and developing technical requirements from business goals and needs. [Required] Experience in solution integration and operability. [Required] Experience working more »
No SQL databases BI tools (Tableau, Power BI etc.) Federated query tools such as Presto/Trino Data lake file formats such as Avro, Parquet, ORC Responsibilities Design the data architecture of organization to support data driven vision. Create design and blueprint of the data capabilities for the organization more »
No SQL databases BI tools (Tableau, Power BI etc.) Federated query tools such as Presto/Trino Data lake file formats such as Avro, Parquet, ORC Responsibilities Design the data architecture of organization to support data driven vision. Create design and blueprint of the data capabilities for the organization more »
new technologies and frameworks Nice to have: Knowledge of databases, SQL Familiarity with Boost ASIO Familiarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, Flatbuffers Experience with gRPC, http/REST and Websocket protocols Experience with Google Cloud/AWS and/or containerization in more »
of data modelling (particularly star-schema) and can produce, maintain, and update relevant data models for specific business needs including bespoke serialization techniques (e.g. Parquet) and table formats (Delta etc) This is an incredibly exciting role that rarely becomes available. You will own a well-designed data lake serving more »
other key languages (especially Python). Expert in key data engineering platforms such as Kafka or other streaming technologies, data lakes (AWS S3, Iceberg, Parquet), analytics technologies (Trinio, Spark), automation technologies (Airflow, ML Flow) and data governance (DataHub). People management and technical leadership experience. Are passionate about agile … IAM, EC2, S3, DynamoDB, MSK). Our Technology Stack ð» Scala and Python Kafka, Spark, Kafka Streams, Kinesis, Akka and KSQL AWS, S3, Iceberg, Parquet, Glue and Spark/EMR for our Data Lake Elasticsearch, Dynamodb and Redis Starburst and Athena Airflow and ML Flow Additional InformationWhy should you more »
is redeveloping our teams data platform from a legacy system to one based on AWS EMR, with Amazon RDS and DynamoDB ingestion converted to Parquet files, interrogatable through Spark and MapReduce. This modern platform will support rapid data insight generation, data experiments for new product development, our live Machine … Amazon DynamoDB, AWS Glue, MapReduce, Hive, Spark, YARN, Airflow. Ability to work with a range of structured, semi-structured and unstructured file formats including Parquet, json, csv, pdf, jpg. Accomplished data wrangling and scripting abilities, with advanced knowledge of Python, Linux shell scripting, SQL and NoSQL. Tools and methods more »
ROLE SUMMARY DATA ENGINEER REMOTE/TRAVEL TO CLIENT SITE IN LONDON Valcon UK are looking to recruit a self-motivated, highly logical and intellectually curious Data Engineer’s to join our expanding Data capability in the UK. As a more »
Chicago, Illinois, United States Hybrid / WFH Options
Request Technology - Robyn Honquest
data messaging design of data science data analytics Kafka and protocol buffers SQL no SQL tableau power bi presto/trino data lakes avro parquet ORC infrastructure technologies ServiceNow or similar. 10 years as a senior data architect data engineer DBA lead logical and conceptual data models data modelling … tools (Tableau, Power BI etc) Expertise with federated query tools such as Presto/Trino Experience with data lake file formats such as Avro, Parquet, ORC [Required] Experience in extracting and developing technical requirements from business goals and needs. [Required] Experience in solution integration and operability. [Required] Experience working more »