Please note that is 100% remote role - but we are going to consider only candidates from from Warsaw and surroundings
About:
We are seeking a highly motivated and self-driven data engineer for our growing data team -who is able to work and deliver independently and as a team. In this role, you will play a crucial part in designing, building and maintaining our ETL infrastructure and data pipelines.
Responsibilities:
● This position is for a Cloud Data engineer with a background in Python, DBT, SQL and data warehousing for enterprise-level systems.
● Adhere to standard coding principles and standards.
● Build and optimize data pipelines for efficient data ingestion, transformation and loading from various sources while ensuring data quality and integrity.
● Design, develop, and deploy python scripts and ETL processes in ADF environment to process and analyze varying volumes of data.
● Experience of DWH, Data Integration, Cloud, Design and Data Modelling.
● Proficient in developing programs in Python and SQL
● Experience with Data warehouse Dimensional data modeling.
● Working with event based/streaming technologies to ingest and process data.
● Working with structured, semi structured and unstructured data.
● Optimize ETL jobs for performance and scalability to handle big data workloads.
● Monitor and troubleshoot ADF jobs, identify and resolve issues or bottlenecks.
● Implement best practices for data management, security, and governance within the Databricks environment. Experience designing and developing Enterprise Data Warehouse solutions.
● Proficient writing SQL queries and programming including stored procedures and reverse engineering existing process.
● Perform code reviews to ensure fit to requirements, optimal execution patterns and adherence to established standards.
● Checking in, checkout and peer review and merging PRs into git Repo.
● Knowledge of deployment of packages and code migrations to stage and prod environments via CI/CD pipelines.
Requirements
● 3+ years Python coding experience.
● 5+ years - SQL Server based development of large datasets
● 5+ years with Experience with developing and deploying ETL pipelines using Databricks Pyspark.
● Experience in any cloud data warehouse like Synapse, ADF, Redshift, Snowflake.
● Experience in Data warehousing - OLTP, OLAP, Dimensions, Facts, and Data modeling.
● Previous experience leading an enterprise-wide Cloud Data Platform migration with strong architectural and design skills.
● Experience with Cloud based data architectures, messaging, and analytics.
● Cloud certification(s).
● Add ons: Any experience with Airflow , AWS lambda, AWS glue and Step functions is a Plus.
Net per month - B2B
Gross per month - Permanent
Check similar offers
Check similar offers