#1 Job Board for tech industry in Europe

  • Job offers
  • All offersKrakówDataSenior Data Engineer
    Senior Data Engineer
    Data
    SAVENTIC HEALTH sp. z o.o.

    Senior Data Engineer

    SAVENTIC HEALTH sp. z o.o.
    Kraków
    Type of work
    Full-time
    Experience
    Senior
    Employment Type
    B2B
    Operating mode
    Remote

    Tech stack

      Python

      advanced

      Big Data

      advanced

      Docker

      regular

      Terraform

      regular

      Linux

      regular

      Kubernetes

      regular

      SQL

      regular

    Job description

    Online interview
    Friendly offer

    Introduction to Saventic:

    Saventic Health - we are an international scale-up company focusing on innovations in rare diseases. We create algorithms, based on artificial intelligence, to support the diagnosis of rare diseases. We represent a unique team of professionals in medicine, science, technology, and management.


    Currently, we are looking for an experienced Data Engineer to develop our analytical capabilities. We plan to perform the implementation of a new architecture combining data extraction from medical centers, data quality verification, creation of a feature store and upload of the data to the database. Additionally, we aim to develop a platform, one of the functionalities of which will be the detection of clinical symptoms in patients who, do not have specific phrases written in their medical description. In the next stages, we plan to develop a large model that will be the core for finding rare diseases.


    Position that we offer:

    • Development and maintenance of ETL processes, data pipelines in "BIG DATA" architecture
    • Creation and maintenance of data pipelines: receiving data from medical units and structuring data in an internal database
    • Expansion of tools supporting data quality control processes
    • Optimization of data processing processes and SQL query optimization
    • Automation of data delivery processes to the data warehouse
    • Monitoring data delivery and processing processes
    • Creation of Data Lake, Data Mart and Feature Store for visualization and modeling purposes
    • Creation of a pipeline for data processing in the scope of CI/CD


    Required Qualification:

    • MD in Engineering, Computer Science, Math, Statistics or related fields
    • 5+ years of relevant experience (can be mixed industry and academic)
    • Practical experience in working with on-premise environments
    • Work in the "BIG DATA" methodology
    • Experience using: Linux, Git, Python, Terraform, Docker, and Kubernetes
    • Experience tools: Kedro, Soda, Great Expectations, Airflow
    • Experience in designing and creating optimal ETL processes in the Big Data domain
    • Practical knowledge of SQL
    • Good programming skills in Python
    • Commercial experience in implementing ETL processes on-premises
    • Understanding of the data model lifecycle
    • English minimum B2


    Our offer:

    • Start-up culture
    • Internationals projects that make a real impact on patients' lives
    • Challenging job in multidisciplinary environment
    • Benefits package