- 💰Up to $7500 per month
- 📍 Hybrid from Warsaw (3 days/week at office)
- ☑️ Full-time, long-term
- 📑 Contract of Employment
We are seeking a highly motivated and self-driven Data Engineer to join our growing data team. The ideal candidate is capable of delivering independently and collaboratively in a team setting. You will play a key role in designing, building, and maintaining scalable ETL infrastructure and data pipelines to support data-driven decision-making across the organization.
- Design, develop, and maintain efficient data pipelines for ingestion, transformation, and loading from various data sources.
- Implement and optimize ETL processes using Python, DBT, SQL, ADF, and Databricks.
- Build scalable data solutions in cloud-based environments (Azure, AWS, Snowflake, Redshift, etc.).
- Develop and deploy Python-based scripts and workflows in Azure Data Factory (ADF).
- Ensure data quality, integrity, and security in line with enterprise standards.
- Monitor, troubleshoot, and optimize performance of ETL and data pipelines.
- Apply dimensional modeling techniques for Data Warehouse design (OLTP/OLAP, Dimensions, Facts).
- Work with structured, semi-structured, and unstructured data.
- Collaborate with cross-functional teams on enterprise-wide cloud data platform migrations.
- Conduct code reviews and follow best practices in software development (CI/CD, Git, testing).
- Ensure data governance, privacy, and compliance in the Databricks environment.
- Contribute to deployment processes via CI/CD pipelines, including code migration to staging and production.
- 3+ years of experience in Python programming.
- 5+ years of hands-on experience in SQL and large dataset processing (preferably SQL Server)
- 5+ years of experience developing and deploying ETL pipelines using Databricks (PySpark).
- Solid experience with cloud data platforms such as Azure Synapse, ADF, Redshift, or Snowflake.
- Expertise in data warehousing concepts and dimensional data modeling.
- Prior experience with enterprise-scale cloud data migrations.
- Understanding of cloud-native data architectures and streaming technologies.
- Hands-on experience with Git and CI/CD pipelines.
- Cloud certification(s) (e.g., Azure, AWS, or GCP) are strongly preferred.
- Experience with Apache Airflow, AWS Lambda, AWS Glue, or Step Functions.
- Familiarity with reverse engineering legacy ETL processes and performance tuning.
We are a technology consulting company and a recruitment agency, delivering software solutions to clients from Europe and the US. We work 100% remotely, in an international team. We employ people with experience in international corporations as well as the ones from the best technical and business universities.
Find out more: https://devsdata.com