#1 Job Board for tech industry in Europe

Cloud Data Engineer
New
Data

Cloud Data Engineer

Warszawa
4 500 - 7 000 USD/monthNet per month - B2B
4 500 - 7 000 USD/monthNet per month - B2B
Type of work
Full-time
Experience
Mid
Employment Type
B2B
Operating mode
Hybrid

Tech stack

    English

    C1

    ETL

    advanced

    Python

    regular

    SQL

    regular

    Airflow

    regular

    OLAP

    regular

Job description

Online interview
Friendly offer

Data Engineer – Cloud ETL & Platform Development


📍 Hybrid in Warsaw (3 days/week onsite required) | 💼 Full-time | B2B contract up to $7,000/month


about the role


Our client is a US-based technology company headquartered in New York City, delivering digital solutions and consulting services that transform businesses and drive measurable value. With offices in multiple countries, the company is now investing in a new engineering centre in Warsaw, recognizing the strong talent and culture of Polish software professionals.


We are looking for a highly motivated and self-driven Data Engineer to join a fast-growing data platform team. In this role, you will design, build, and maintain robust, scalable, and cloud-native ETL infrastructure and data pipelines, enabling real-time analytics and AI-ready architecture for high-performance business applications.

This is a hybrid position — you will be expected to work from the Warsaw office at least 3 days per week.


Key Responsibilities


  • Design, develop, and deploy Python-based ETL pipelines using Airflow and Prefect

  • Build and optimize data warehouse structures for analytics, OLAP, and dimensional modeling

  • Model AI-compatible DB schemas and ontologies for future-facing analytics

  • Migrate and transform structured, semi-structured, and unstructured data using DBT and Pandas

  • Work with streaming/event-driven architectures for real-time data processing

  • Optimize pipeline performance and scalability for large data volumes

  • Ensure data quality, validation, cleansing, and graceful error handling

  • Perform code reviews to ensure standards, scalability, and best practices

  • Collaborate with DevOps on CI/CD and automated release pipelines

  • Implement data governance, security, and observability across ETL processes


Key Requirements


  • 5+ years of experience designing and building enterprise-scale ETL pipelines

  • Strong Python skills, including use of Pandas for data processing

  • Proficient in SQL — writing complex queries and procedures across RDBMSes

  • Experience with Airflow, Prefect, or similar workflow orchestration tools

  • Solid understanding of data warehousing (OLTP, OLAP, Facts & Dimensions)

  • Familiarity with cloud-based data platforms such as RDS, Redshift, or Snowflake

  • Knowledge of cloud data architectures, messaging systems, and scalable design

  • Hands-on experience with code versioning, deployment automation, and CI/CD


Preferred Qualifications


  • Experience with Docker, Kubernetes, AWS Lambdas, Step Functions

  • Exposure to Databricks, PySpark, and advanced distributed data processing

  • Cloud certifications are a plus


What’s Offered


  • B2B contract with monthly compensation up to $7,000

  • Strong career growth opportunities in a global fintech environment

  • High-impact projects in a fast-growing sector

  • Friendly, open, and ambitious team culture

  • Hybrid model – minimum 3 days/week in the Warsaw office

4 500 - 7 000 USD/month

Net per month - B2B