BigData Developer with Python/Spark/ETL
Requirements
– Professional data engineering experience focused on batch and real-time data pipelines development using Spark and Python
– Data processing/data transformation using ETL tools, Azure Databricks platform (preferred)
– Cloud Data Warehouse solutions experience (Snowflake, Azure DW, or Redshift)
– Proactive approach to problem-solving with effective influencing skills
– Familiar with Agile practices and methodologies
Responsibilities
We are looking for a Data Engineer for the Enterprise Data Organization to build and manage data pipeline (Data ingest, data transformation, data distribution, quality rules, data storage, etc.) for the Azure cloud-based data platform.
Nice-to-Have Experience
– Experience with a DevOps model utilizing a CI/CD tool
– Experienced in Azure Cloud Platform
– Hands-on Talend work experience (anyone with this skill will have an advantage over other candidates)
– Apache Airflow, Azure Data Factory experience
What we offer
– Opportunity to work on bleeding-edge projects
– Work with a highly motivated and dedicated team
– Competitive salary
– Flexible schedule
– Benefits program
– Social package – medical insurance, sports
– Corporate social events
– Professional development opportunities
– Opportunity for long business trips to the US and possibility for relocation