BigData Developer with Python/Spark/ETL

Необхідні навички

– Professional data engineering experience focused on batch and real-time data pipelines development using Spark and Python
– Data processing/data transformation using ETL tools, Azure Databricks platform (preferred)
– Cloud Data Warehouse solutions experience (Snowflake, Azure DW, or Redshift)
– Proactive approach to problem-solving with effective influencing skills
– Familiar with Agile practices and methodologies


We are looking for a Data Engineer for the Enterprise Data Organization to build and manage data pipeline (Data ingest, data transformation, data distribution, quality rules, data storage, etc.) for the Azure cloud-based data platform.

Буде плюсом

– Experience with a DevOps model utilizing a CI/CD tool
– Experienced in Azure Cloud Platform
– Hands-on Talend work experience (anyone with this skill will have an advantage over other candidates)
– Apache Airflow, Azure Data Factory experience


– Opportunity to work on bleeding-edge projects
– Work with a highly motivated and dedicated team
– Competitive salary
– Flexible schedule
– Benefits program
– Social package – medical insurance, sports
– Corporate social events
– Professional development opportunities
– Opportunity for long business trips to the US and possibility for relocation


    Локація: Київ, Львів, Харьків
    Зайнятість: повна

    Our partner is the engineering services company known for transformative, mission-critical cloud solutions for retail, finance and technology sectors. We architected some of the busiest e-commerce services on the Internet and have never had an outage during the peak season.