#1 Job Board for tech industry in Europe

  • Job offers
  • Data Engineer with Databricks
    New
    Data

    Data Engineer with Databricks

    Warszawa
    5 926 - 6 350 USDNet/month - B2B
    Type of work
    Full-time
    Experience
    Mid
    Employment Type
    B2B
    Operating mode
    Remote

    Tech stack

      Databricks

      regular

      Python

      regular

      Spark

      regular

      SQL

      regular

      Azure

      regular

    Job description

    Online interview

    Responsibilities:

    • Storage Cleanup
    • Implement cost tracking at user-level
    • Transition to serverless compute for optimized performance
    • Move scheduled jobs from all-purpose clusters to dedicated job clusters
    • Proactively monitor costs and performance with dashboards and alerts
    • Ofload raw data to DWH landing zone
    • Use pre-aggregated datasets to reduce duplicate compute
    • Transition Databricks workloads to DBT framework
    • Complete migration of scheduled jobs to dedicated job clusters
    • Expand serverless compute deployment post-GA

     

    Requirements:

    • Minimum 3 years of experience in similar position
    • Proficiency in Spark, with a focus on Scala and Python
    • Strong experience with Azure, including Azure Data Factory (ADF) and Azure Storage
    • Hands-on expertise with Azure Databricks, including IAM, Jobs, Workflows, SQL Warehouses, Delta Lake, and Serverless
    • Advanced SQL skills for data transformation and optimization
    • Experience with DBT (Data Build Tool) for data modeling and pipeline development
    • English level: B2, C1

    Nice to have:

    • Experience with Kafka for real-time data streaming

     

    Responsibilities:

    • Storage Cleanup
    • Implement cost tracking at user-level
    • Transition to serverless compute for optimized performance
    • Move scheduled jobs from all-purpose clusters to dedicated job clusters
    • Proactively monitor costs and performance with dashboards and alerts
    • Ofload raw data to DWH landing zone
    • Use pre-aggregated datasets to reduce duplicate compute
    • Transition Databricks workloads to DBT framework
    • Complete migration of scheduled jobs to dedicated job clusters
    • Expand serverless compute deployment post-GA

     

    Requirements:

    • Minimum 3 years of experience in similar position
    • Proficiency in Spark, with a focus on Scala and Python
    • Strong experience with Azure, including Azure Data Factory (ADF) and Azure Storage
    • Hands-on expertise with Azure Databricks, including IAM, Jobs, Workflows, SQL Warehouses, Delta Lake, and Serverless
    • Advanced SQL skills for data transformation and optimization
    • Experience with DBT (Data Build Tool) for data modeling and pipeline development
    • English level: B2, C1

    Nice to have:

    • Experience with Kafka for real-time data streaming

     

    Offer:

    • Private medical care
    • Co-financing for the sports card
    • Training & learning opportunities
    • Constant support of dedicated consultant
    • Employee referral program


    tutlo_banner_hero

    Practice your English before your job interview!

    Get 3 free English lessons
    5 926 - 6 350 USD

    Net/month - B2B

    Check similar offers

    Data Scientist

    New
    DCG
    4.79K - 5.29K USD
    Warszawa
    , Fully remote
    Fully remote
    Python
    Time series modeling
    English

    Power BI Developer

    New
    Aspire Systems Poland
    3.53K - 6.05K USD
    Gdańsk
    , Fully remote
    Fully remote
    Power BI
    Data Visualization
    AWS

    Data Engineer (Python, Databricks)

    New
    emagine Polska
    6.75K - 7.81K USD
    Wrocław
    , Fully remote
    Fully remote
    Python
    Azure Databricks
    CI/CD

    Analityk/Analityczka Systemowy/a

    New
    Nest Bank S.A.
    4.23K - 5.5K USD
    Kraków
    , Fully remote
    Fully remote
    English
    UML
    BPMN

    Data Science Engineer

    New
    HAYS Poland
    5.07K - 5.49K USD
    Kraków
    , Fully remote
    Fully remote
    ML models
    BigQuery
    GCP