Reading, Berkshire, United Kingdom Hybrid / WFH Options
Applicable Limited
using S3Input and S3Output components. Handle large-scale data sets and optimize ETL processes for performance and scalability. Integrate Talend workflows with data lakes, employing file formats such as Parquet, ORC, or Avro for efficient storage and retrieval. Work with NoSQL databases for specific use cases, ensuring seamless integration with Talend workflows. Leverage Apache Spark for data processing and … is highly preferred. Experience with cloud platforms, especially AWS S3, and knowledge of data lake architectures. Familiarity with big data tools such as Apache Spark and file formats like Parquet, ORC, and Avro. Experience with NoSQL databases and their integration with Talend. Solid understanding of CI/CD pipelines and YAML configuration in DevOps contexts. Familiarity with Amazon Glue More ❯
South East London, London, United Kingdom Hybrid / WFH Options
Datatech Analytics
processing and automation Solid understanding of ETL/ELT workflows, data modelling, and structuring datasets for analytics Experience working with large, complex datasets and APIs across formats (CSV, JSON, Parquet, etc) Familiarity with workflow automation tools (eg, Power Automate) and/or Power Apps is desirable Excellent interpersonal and communication skills with the ability to work cross-functionally and More ❯
Nursling, Southampton, Hampshire, England, United Kingdom Hybrid / WFH Options
Ordnance Survey
Survey Testing Community, with common standards such as metrics and use of test tools Here is a snapshot of the technologies that we use Scala, Apache Spark, Databricks, ApacheParquet, YAML, Azure Cloud Platform, Azure DevOps (Test plans, Backlogs, Pipelines), GIT, GeoJSON What we're looking for Highly skilled in creating, maintaining and peer reviewing test automation code, preferably More ❯
About the role Taktile is a high-growth, post product-market-fit start-up, on a fast trajectory to becoming market leader in the field of automated decisioning. We are looking for a Full-stack Engineer to join the Decide More ❯
and RAG-based solutions Proficiency in Python and modern AI/ML libraries (e.g. HuggingFace, LangChain, TensorFlow, PyTorch) Experience with data exchange and storage frameworks (e.g. APIs, SQL, NoSQL, Parquet) Track record of delivering technical solutions in Agile environments Excellent communication skills and a collaborative mindset Beneficial, but not essential: Experience with containerisation (Docker) Awareness of secure data handling More ❯
Bonus Points For Workflow orchestration tools like Airflow. Working knowledge of Kafka and Kafka Connect. Experience with Delta Lake and lakehouse architectures. Proficiency in data serialization formats: JSON, XML, PARQUET, YAML. Cloud-based data services experience. Ready to build the future of data? If you're a collaborative, forward-thinking engineer who wants to work on meaningful, complex problems More ❯