development experience in at least one of the following: Java, Python, PySpark or Scala.Experience working with a variety of data formats such as JSON, Parquet, XML etc.Experience with or developed understanding of the application of ETL/ELT tools.Experience with NoSQL type environments, Data Lakes, Lake-Houses (Cassandra, MongoDB more »
flows & pipelines in a cloud environment using a progressive tech stack e.g. Databricks, Spark, Python, PySpark, Delta Lake, SQL, Logic Apps, Azure Functions, ADLS, Parquet, Neo4J, Flask Ingest and integrate data from a large number of disparate data sources Design and build complex data models for analytical insight Write more »
relational models, and formats to support geo-enabled applications and graph database structures, using knowledge of various data formats including XML, JSON, flat file, Parquet etc. You can also manage metadata across your development teams systems and beyond, with a practitioner knowledge of tracking data provenance and data lineage. more »
Ensure systems meet business requirements and industry practices for data integrity and quality. Manage ETL and ELT pipelines across many data sources (CSV/parquet files, API endpoints, etc) Design and build data models for the business end users. Write complex SQL queries for standard as well as ad more »
/OpenShift including high availability Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
scale data processing and analytics Desired experience: Worked with Python 3.9+ Familiar with Python test automation Experience with SQL and Timeseries databases Familiar with Parquet, Arrow, Airflow, Databricks Experience with cloud AWS services, such as S3, EC2, RDS etc Quality engineering best practice and tooling including TDD, BDD This more »
new technologies and frameworks Nice to have: Knowledge of databases, SQL Familiarity with Boost ASIO Familiarity with data serialization formats such Apache Arrow/Parquet, Google Protocol Buffers, Flatbuffers Experience with gRPC, http/REST and Websocket protocols Experience with Google Cloud/AWS and/or containerization in more »
of data modelling (particularly star-schema) and can produce, maintain, and update relevant data models for specific business needs including bespoke serialization techniques (e.g. Parquet) and table formats (Delta etc) This is an incredibly exciting role that rarely becomes available. You will own a well-designed data lake serving more »
other key languages (especially Python). Expert in key data engineering platforms such as Kafka or other streaming technologies, data lakes (AWS S3, Iceberg, Parquet), analytics technologies (Trinio, Spark), automation technologies (Airflow, ML Flow) and data governance (DataHub). People management and technical leadership experience. Are passionate about agile … IAM, EC2, S3, DynamoDB, MSK). Our Technology Stack ð» Scala and Python Kafka, Spark, Kafka Streams, Kinesis, Akka and KSQL AWS, S3, Iceberg, Parquet, Glue and Spark/EMR for our Data Lake Elasticsearch, Dynamodb and Redis Starburst and Athena Airflow and ML Flow Additional InformationWhy should you more »
ROLE SUMMARY DATA ENGINEER REMOTE/TRAVEL TO CLIENT SITE IN LONDON Valcon UK are looking to recruit a self-motivated, highly logical and intellectually curious Data Engineer’s to join our expanding Data capability in the UK. As a more »