#1 Job Board for tech industry in Europe

  • Job offers
  • Data Engineer
    Data

    Data Engineer

    Szczecin
    Type of work
    Full-time
    Experience
    Mid
    Employment Type
    B2B
    Operating mode
    Remote

    Tech stack

      Python

      advanced

      Big Data

      advanced

      Docker

      regular

      Terraform

      regular

      Linux

      regular

      Kubernetes

      regular

      SQL

      regular

    Job description

    Online interview
    Friendly offer

    Introduction to Saventic:

    Saventic Health - we are an international scale-up company focusing on innovations in rare diseases. We create algorithms, based on artificial intelligence, to support the diagnosis of rare diseases. We represent a unique team of professionals in medicine, science, technology, and management.


    Currently, we are looking for Data Engineer to develop our analytical capabilities. We plan to perform the implementation of a new architecture combining data extraction from medical centers, data quality verification, creation of a feature store and upload of the data to the database. Additionally, we aim to develop a platform, one of the functionalities of which will be the detection of clinical symptoms in patients who, do not have specific phrases written in their medical description. In the next stages, we plan to develop a large model that will be the core for finding rare diseases.


    Position that we offer:

    • Development and maintenance of ETL processes, data pipelines in "BIG DATA" architecture
    • Creation and maintenance of data pipelines: receiving data from medical units and structuring data in an internal database
    • Expansion of tools supporting data quality control processes
    • Optimization of data processing processes and SQL query optimization
    • Automation of data delivery processes to the data warehouse
    • Monitoring data delivery and processing processes
    • Creation of Data Lake, Data Mart and Feature Store for visualization and modeling purposes
    • Creation of a pipeline for data processing in the scope of CI/CD


    Required Qualification:

    • MD in Engineering, Computer Science, Math, Statistics or related fields
    • 3+ years of relevant experience
    • Practical experience in working with on-premise environments will be advantage
    • Work in the "BIG DATA" methodology
    • Experience using: Linux, Git, Python, Terraform, Docker, and Kubernetes
    • Experience tools: Kedro, Soda, Great Expectations, Airflow
    • Experience in designing and creating optimal ETL processes in the Big Data domain
    • Practical knowledge of SQL
    • Good programming skills in Python
    • Commercial experience in implementing ETL processes on-premises will be advantage
    • Understanding of the data model lifecycle
    • English minimum B2 (spanish or any others will be +)
    • Open to traveling several times a year


    Our offer:

    • Start-up culture
    • Internationals projects that make a real impact on patients' lives
    • Challenging job in multidisciplinary and international environment
    • Benefits package


    Check similar offers

    Product Data Analyst

    New
    Booksy
    Undisclosed Salary
    Warszawa
    , Fully remote
    Fully remote
    SQL
    BigQuery
    PowerBi

    Programista APEX

    New
    Aplikacje Krytyczne
    3.1K - 3.84K USD
    Warszawa
    , Fully remote
    Fully remote
    Oracle
    PL/SQL
    REST API

    Data Managment Specialist

    New
    ADASTRA
    5.36K - 6.97K USD
    Warszawa
    , Fully remote
    Fully remote
    Data Managment
    data governance
    Data Quality

    Data Software Engineer (Upskilling position for Python Developers)

    New
    EPAM Systems
    Undisclosed Salary
    Szczecin
    , Fully remote
    Fully remote
    AWS
    GCP
    Python

    Oracle Developer

    New
    FinDev
    3.8K - 4.8K USD
    Warszawa
    , Fully remote
    Fully remote
    Oracle SQL
    PL/SQL
    Python