#1 Job Board for tech industry in Europe

Data Engineer – Common Data Layer / Data Hubs / GenAI
New
Data

Data Engineer – Common Data Layer / Data Hubs / GenAI

44 USD/hNet per hour - B2B
44 USD/hNet per hour - B2B
Type of work
Full-time
Experience
Mid
Employment Type
B2B
Operating mode
Hybrid

Tech stack

    English

    B2

    Polish

    C1

    Machine Learning

    regular

    CI/CD

    regular

    ETL

    regular

    Azure SQL

    regular

    AI

    regular

    ELT

    regular

    Azure Data Factory

    regular

    Terraform

    regular

    ADLS

    regular

    SQL

    regular

Job description

Project information:

  • Industry: insurance and IT services
  • Rate: up to 160 PLN/H net + VAT, B2B
  • Location: Warsaw – first 2-3 months hybrid model of work, then remote
  • Project language: Polish, English

 

Summary

As a Data Engineer, the primary function of this role is to design, build, and maintain Data Hubs that integrate data from multiple sources to support analytical, reporting, operational, and Generative AI use cases. The role is essential to create a scalable and efficient data infrastructure that facilitates real-time data availability and model updates. Collaboration is expected with data architects, AI engineers, and business teams. Tools include Databricks, Azure Data Factory, and Azure SQL.


Responsibilities:

  • Data Hub Development – Design and implement scalable Data Hubs to support enterprise-wide data needs.
  • Data Pipeline Engineering – Build and optimize ETL/ELT pipelines for efficient data ingestion, transformation, and storage.
  • Logical Data Modeling – Structure Data Hubs to ensure efficient access patterns and support diverse use cases.
  • Real-Time Analytics – Enable real-time data ingestion and updating models to support streaming and real-time analytics.
  • Data Quality & Monitoring – Develop data validation, anomaly detection, and monitoring features to ensure high data reliability.
  • Performance Optimization – Optimize data processing and storage for large-scale datasets.
  • Automation & CI/CD – Implement CI/CD pipelines to automate data workflows and deployments.
  • Collaboration – Work with data architects, AI engineers, and business teams to align data solutions.
  • Monitoring & Maintenance – Continuously improve data infrastructure for scalability and reliability.
  • Agile Practices – Participate in Scrum/Agile methodologies to deliver high-quality data solutions.
  • Documentation – Create and maintain clear, structured documentation for data models, pipelines, and technical decisions.


Key Requirements (4-7 years of experience)

  • Strong Python skills for data engineering.
  • Experience with Azure Data Factory, ADLS, and Azure SQL.
  • Hands-on experience in ETL/ELT development.
  • Experience with real-time data processing.
  • Understanding of AI/ML data processing.
  • Proficiency in SQL.
  • Knowledge of CI/CD and infrastructure-as-code (Terraform).
  • Understanding of data governance and compliance.
  • Experience with Databricks and Apache Spark.
  • Familiarity with containerization (Docker, Kubernetes).
  • Ability to produce high-quality technical documentation.
  • Minimum B2 level in English.


Nice to have:

  • Background from large consulting firms.
  • Graduation from prestigious universities.


44 USD/h

Net per hour - B2B

Apply for this job

File upload
Add document

Format: PDF, DOCX, JPEG, PNG. Max size 5 MB

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Informujemy, że administratorem danych jest emagine z siedzibą w Warszawie, ul.Domaniewskiej 39A (dalej jako "administra... more