#1 Job Board for tech industry in Europe

  • Job offers
  • Data Engineer – Common Data Layer / Data Hubs / GenAI
    New
    Data

    Data Engineer – Common Data Layer / Data Hubs / GenAI

    Warszawa
    42 USD/hNet per hour - B2B
    42 USD/hNet per hour - B2B
    Type of work
    Full-time
    Experience
    Mid
    Employment Type
    B2B
    Operating mode
    Hybrid

    Tech stack

      English

      B2

      Polish

      C1

      Machine Learning

      regular

      CI/CD

      regular

      ETL

      regular

      Azure SQL

      regular

      AI

      regular

      ELT

      regular

      Azure Data Factory

      regular

      Terraform

      regular

      ADLS

      regular

      SQL

      regular

    Job description

    Project information:

    • Industry: insurance and IT services
    • Rate: up to 160 PLN/H net + VAT, B2B
    • Location: Warsaw – first 2-3 months hybrid model of work, then remote
    • Project language: Polish, English

     

    Summary

    As a Data Engineer, the primary function of this role is to design, build, and maintain Data Hubs that integrate data from multiple sources to support analytical, reporting, operational, and Generative AI use cases. The role is essential to create a scalable and efficient data infrastructure that facilitates real-time data availability and model updates. Collaboration is expected with data architects, AI engineers, and business teams. Tools include Databricks, Azure Data Factory, and Azure SQL.


    Responsibilities:

    • Data Hub Development – Design and implement scalable Data Hubs to support enterprise-wide data needs.
    • Data Pipeline Engineering – Build and optimize ETL/ELT pipelines for efficient data ingestion, transformation, and storage.
    • Logical Data Modeling – Structure Data Hubs to ensure efficient access patterns and support diverse use cases.
    • Real-Time Analytics – Enable real-time data ingestion and updating models to support streaming and real-time analytics.
    • Data Quality & Monitoring – Develop data validation, anomaly detection, and monitoring features to ensure high data reliability.
    • Performance Optimization – Optimize data processing and storage for large-scale datasets.
    • Automation & CI/CD – Implement CI/CD pipelines to automate data workflows and deployments.
    • Collaboration – Work with data architects, AI engineers, and business teams to align data solutions.
    • Monitoring & Maintenance – Continuously improve data infrastructure for scalability and reliability.
    • Agile Practices – Participate in Scrum/Agile methodologies to deliver high-quality data solutions.
    • Documentation – Create and maintain clear, structured documentation for data models, pipelines, and technical decisions.


    Key Requirements (4-7 years of experience)

    • Strong Python skills for data engineering.
    • Experience with Azure Data Factory, ADLS, and Azure SQL.
    • Hands-on experience in ETL/ELT development.
    • Experience with real-time data processing.
    • Understanding of AI/ML data processing.
    • Proficiency in SQL.
    • Knowledge of CI/CD and infrastructure-as-code (Terraform).
    • Understanding of data governance and compliance.
    • Experience with Databricks and Apache Spark.
    • Familiarity with containerization (Docker, Kubernetes).
    • Ability to produce high-quality technical documentation.
    • Minimum B2 level in English.


    Nice to have:

    • Background from large consulting firms.
    • Graduation from prestigious universities.


    42 USD/h

    Net per hour - B2B

    Apply for this job

    File upload
    Add document

    Format: PDF, DOCX, JPEG, PNG. Max size 5 MB

    This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
    Informujemy, że administratorem danych jest emagine z siedzibą w Warszawie, ul.Domaniewskiej 39A (dalej jako "administra...more

    Check similar offers

    Data Engineer (AWS)

    New
    Addepto
    4.02K - 5.59K USD/month
    Warszawa
    , Fully remote
    Fully remote
    Spark
    AWS
    SQL