Shawnee Mission, Kansas, United States Hybrid / WFH Options
ECCO Select
services and integrations using Java (primary) and optionally Python for ETL workflows and data transformation. • Integrate data from internal and external REST APIs; handle data format translation (e.g., JSON, Parquet, Avro). • Optimize data flows for reliability and performance, and support large-scale batch and streaming data jobs. • Implement and document ETL mappings, schemas, and transformation logic aligned with More ❯
Shawnee Mission, Kansas, United States Hybrid / WFH Options
ECCO Select
services and integrations using Java (primary) and optionally Python for ETL workflows and data transformation. - Integrate data from internal and external REST APIs; handle data format translation (e.g., JSON, Parquet, Avro). - Optimize data flows for reliability and performance, and support large-scale batch and streaming data jobs. - Implement and document ETL mappings, schemas, and transformation logic aligned with More ❯
Excellent stakeholder engagement and communication skills — confident working cross-functionally and enabling others to use and trust the data Desirable: Experience with Lakehouse patterns; setting standards, quality checks etc. Parquet partitioning and clustering to optimise for query speed and cost GitHub Actions REST APIs ML Ops tooling What you can expect 25 days holiday excluding public holidays increasing after More ❯
Birmingham, West Midlands, United Kingdom Hybrid / WFH Options
Amtis Professional Ltd
continuously improve data infrastructure Explore AI-driven enhancements to boost data accuracy and productivity Requirements: Strong experience with: Azure Databricks, Data Factory, Blob Storage Python/PySpark SQL Server, Parquet, Delta Lake Deep understanding of: ETL/ELT, CDC, stream processing Lakehouse architecture and data warehousing Scalable pipeline design and database optimisation A proactive mindset, strong problem-solving skills More ❯
Hampton, Virginia, United States Hybrid / WFH Options
Iron EagleX, Inc
storage solutions and OpenSearch for search functionality. Proficiency in developing and managing APIs using FastAPI. Experience using Trino for query optimization and distributed SQL processing. Proficiency in working with Parquet Files for efficient data storage and retrieval. Hands-on experience with Airflow for orchestrating complex data workflows. Familiarity with Databricks Medallion model. Familiarity with OpenMetaData for metadata management and More ❯
others in a high-impact role based in the Kansas City Metropolitan Area. Responsibilities: Architect and develop Python-based microservices (FastAPI, Flask, or custom). Translate data formats (JSON, Parquet, Avro) and develop automation/scripting solutions. Drive root cause analysis and troubleshooting across staging and production. Lead integration efforts with DevOps, security, and cloud infrastructure teams. Guide CI More ❯
Data Lake and create secure, efficient, and scalable environments for our data platforms. Leveraging cloud-native technologies and AWS tools such as AWS S3, EKS, Glue, Airflow, Trino, and Parquet, you will prepare to adopt Apache Iceberg for greater performance and flexibility. You'll address high-performance data workloads, ensuring seamless execution of massive queries, including 600+ billion-row More ❯
for continuous deployment AWS cloud infrastructure Kubernetes for data services and task orchestration Google Analytics, Amplitude and Firebase for client applications event processing Airflow for job scheduling and tracking Parquet and Delta file formats on S3 for data lake storage Streamlit for data applications Why else you'll love it here Wondering what the salary for this role is More ❯
SQL, and orchestration tools like Apache Airflow. Hands-on experience with data serialization formats such as protobuf, Arrow, FlatBuffers, or Cap'n Proto. Familiarity with data storage formats like Parquet or Avro. Experience with modern analytic storage technologies such as Apache Iceberg or DuckDB. Binary message parsing experience. Strong understanding of classified data handling, secure networking, and compliance in More ❯
market data feeds (Bloomberg, Refinitiv, etc). Familiarity with containerisation (Docker) and cloud platforms (AWS/GCP). Database experience (SQL/NoSQL) and use of modern data formats (Parquet, HDF5). Contract: Initial term with possible extension Day Rate: Competitive, depending on experience McGregor Boyall is an equal opportunity employer and do not discriminate on any grounds. More ❯
with REST APIs Experience with Java Experience with full lifecycle agile software development projects Desired skills: Experience with Python. Experience building data products in Apache Avro and/or Parquet On-the-job experience with Java software development. Experience deploying the complete DevOps Lifecycle including integration of build pipelines, automated deployments, and compliance scanning using test driven development. Responsibilities More ❯
such as GitHub Actions or Jenkins. Solid grounding in modern engineering principles and full-stack development. Bonus Skills: Airflow, Kafka/Kafka Connect, Delta Lake, JSON/XML/Parquet/YAML, cloud-based data services. Why Apply? Work for a global payments innovator shaping the future of commerce. Join a highly skilled, collaborative, and forward-thinking data team. More ❯
such as GitHub Actions or Jenkins. Solid grounding in modern engineering principles and full-stack development. Bonus Skills: Airflow, Kafka/Kafka Connect, Delta Lake, JSON/XML/Parquet/YAML, cloud-based data services. Why Apply? Work for a global payments innovator shaping the future of commerce. Join a highly skilled, collaborative, and forward-thinking data team. More ❯
as SQL Server, PostgreSQL, Teradata and others. • Proficiency in technologies in the Apache Hadoop ecosystem, especially Hive, Impala and Ranger • Experience working with open file and table formats such Parquet, AVRO, ORC, Iceberg and Delta Lake • Extensive knowledge of automation and software development tools and methodologies. • Excellent working knowledge of Linux. Good working networking knowledge. • Ability to gain customer More ❯
an expert BS degree in Computer Science or meaningful relevant work experience Preferred Qualifications Experience with large scale data platform infrastructure such as Spark, Flink, HDFS, AWS/S3, Parquet, Kubernetes is a plus More ❯
in peer reviews. Collaborate with stakeholders and researchers to support analytics and product development. Integrate data from APIs, S3 buckets, and structured/unstructured sources (JSON, CSV, Excel, PDF, Parquet). Join geospatial datasets with external data sources and apply complex transformations. Define validated data schemas and create clear documentation for partners and teams. Explore and evaluate new data More ❯
in peer reviews. Collaborate with stakeholders and researchers to support analytics and product development. Integrate data from APIs, S3 buckets, and structured/unstructured sources (JSON, CSV, Excel, PDF, Parquet). Join geospatial datasets with external data sources and apply complex transformations. Define validated data schemas and create clear documentation for partners and teams. Explore and evaluate new data More ❯
time for your personal development What you'll be working with: •Backend: Distributed, event-driven core Java (90% of the code-base), MySQL, Kafka •Data analytics: Python & Jupyter notebooks, Parquet, Docker •Testing: JUnit, JMH, JCStress, Jenkins, Selenium, many in-house tools •OS: Linux (Fedora for development, Rocky in production) The LMAX way is to use the right tool for More ❯
bachelor's degree in management information systems, computer science, computer engineering, or similar Strong understanding in modeling structures and unstructured data, and data structures such as Relational, XML, JSON, Parquet, etc. is a plus Strong leadership abilities and demonstrated analytical and problem-solving skills Excellent organization, communication, data analysis and follow-up skills Knowledge of SQL, Python, Java is More ❯
Location London, United Kingdom Employment Type Full time Location Type Remote Department R&D Investigations The engineering team at Chainalysis is inspired by solving the hardest technical challenges and creating products that build trust in cryptocurrencies. We're a global More ❯