GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GCP. + Minimum of 3 years of building and operationalizing large-scale enterprise data solutions using one or more third-party resources such as Pyspark, Talend, Matellion, Informatica or native utilities as Spark, Hive, Cloud DataProc, Cloud Dataflow, Apache, Beam Composer, Big Table, Cloud BigQuery, Cloud PubSub etc. + more »
GBMs, Elastic Net GLMs, GAMs, Decision Trees, Random Forests, Neural Nets and Clustering Experience in statistical and data science programming languages (e.g. R, Python, PySpark, SAS, SQL) A good quantitative degree (Mathematics, Statistics, Engineering, Physics, Computer Science, Actuarial Science) Experience of WTW’s Radar software is preferred Proficient at more »
West 19th Street (22008), United States of America, New York, New York Senior Software Engineer, (Python, PySpark, Docker, Kubernetes, Jenkins) Do you love building and pioneering in the technology space? Do you enjoy solving complex business problems in a fast-paced, collaborative, inclusive, and iterative delivery environment? At Capital more »
AB12, Charlestown, Aberdeen City, City of Aberdeen, United Kingdom
KBM Resourcing
knowledge of API integration. Experience managing CRM systems like Salesforce. Desired Skills: Knowledge of Salesforce development (Apex, Visualforce). Familiarity with Microsoft Fabric and PySpark (Python). Experience with AI tools and models for digital transformation. Understanding of IoT solutions. Personal Attributes: Strong multitasking and teamwork skills. High level more »
Employment Type: Permanent
Salary: £35000 - £55000/annum DOE + Additional benefits
Ability to write code and functions in a way that is reusable, fits best standards guidance, and are robustly unit tested. (Preferably in Python, PySpark or R) Experience of creating reproducible pipelines for systematic processing of statistical data, based on initial design documents. This includes writing documentation, designing and more »
Employment Type: Contract
Rate: £20 - £20.47/hour Flexible Working, Hybrid Working
expertise in data validation for accuracy, completeness, and integrity. JMeter : Experience in API testing using JMeter (Load and Performance testing experience not required). PySpark : Skilled in scripting for data processing using PySpark. Data Lakes : Hands-on experience with data lake environments. Kafka and StreamSets : Familiarity with Kafka StreamSets more »
Charlotte, North Carolina, United States Hybrid / WFH Options
APN Consulting Inc
APN Consulting has an immediate need for a direct client requirement: Python Developer - Pyspark Location - Charlotte, NC (Hybrid) Long-term Contract Key Responsibilities: She was involved in data manipulation using Python scripts Spark Scala which will be useful for faster data processing. Created Parameterized Queries generated Tabular reports Sub more »
The Data Management Specialist will be responsible for managing and optimizing our organization's data infrastructure, ensuring data quality, and implementing data solutions using PySpark, DataBricks, Snowflake, and/or Redshift. The ideal candidate will have a strong background in data management, programming, and cloud-based data platforms. Key … Responsibilities: 1. Develop and maintain data pipelines using PySpark to process and transform large volumes of data. 2. Design, implement, and optimize data solutions using DataBricks for data analytics and machine learning applications. 3. Manage and administer cloud-based data platforms such as Snowflake and Redshift, ensuring high availability … degree preferred. Minimum of 5 years of experience in data management, data engineering, or related role. Strong programming skills in Python and experience with PySpark for data processing and analytics. Hands-on experience with DataBricks for building and optimizing data pipelines. Proficiency in managing and administering cloud-based data more »
Leeds, England, United Kingdom Hybrid / WFH Options
Mastek
OBIEE, Workato and PL/SQL. Design and build data solutions on Azure, leveraging Databricks, Data Factory, and other Azure services. Utilize Python and PySpark for data transformation, analysis, and real-time streaming. Collaborate with cross-functional teams to gather requirements, design solutions, and deliver insights. Implement and maintain … Technologies: Databricks, Data Factory: Expertise in data engineering and orchestration. DevOps, Storage Explorer, Data Studio: Competence in deployment, storage management, and development tools. Python, PySpark: Advanced coding skills, including real-time data streaming through Autoloader. Development Tools: VS Code, Jira, Confluence, Bitbucket. Service Management: Experience with ServiceNow. API Integration more »
Dubai? Senior Data Engineer – Relocate to Dubai! – Full Time – Hybrid Working Are you a skilled Senior Data Engineer with expertise in Python, SQL, and PySpark ? Are you excited to work in a fast-paced financial sector with opportunities for growth and cutting-edge projects? This is your chance to … high-impact data projects and opportunities for professional advancement. What We’re Looking For: Experience: Data Engineer, with proven expertise in Python, SQL , and PySpark . Technical Skills: Strong experience in building and maintaining data pipelines and ETL processes. Familiarity with cloud-based environments (AWS, Azure, GCP) is a more »
Lead Data Engineer: We need some strong Lead data engineer profiles… they need good experience with Python, SQL, ADF and preferably Azure Databricks experience Job description: Building new data pipelines and optimizing data flows using the Azure cloud stack. Building more »
scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Hive SQL, Spark SQL, Python, PySpark and Bash Experience of working with a wide variety of structured and unstructured data. You and your role As a Data Engineer at DWP … you'll be executing code across a full tech stack, including Azure, Databricks, PySpark and Pandas, helping the department to move towards a cloud computing environment, working with huge data sets as part of our DataWorks platform - a system that provides Universal Credit data to our Data Science team more »
scalable, automated ETL pipelines in an AWS cloud environment using AWS S3 Cloud Object Storage Strong coding skills using Hive SQL, Spark SQL, Python, PySpark and Bash Experience of working with a wide variety of structured and unstructured data. You and your role As a Data Engineer at DWP … you'll be executing code across a full tech stack, including Azure, Databricks, PySpark and Pandas, helping the department to move towards a cloud computing environment, working with huge data sets as part of our DataWorks platform - a system that provides Universal Credit data to our Data Science team more »