#1 Job Board for tech industry in Europe

  • Job offers
  • All offersWarszawaDataData Engineer - Data Science Hub
    Data Engineer - Data Science Hub
    Data
    Allegro

    Data Engineer - Data Science Hub

    Allegro
    Warszawa
    Type of work
    Full-time
    Experience
    Mid
    Employment Type
    Permanent
    Operating mode
    Hybrid
    Allegro

    Allegro

    We’re Poland’s most popular shopping platform and the largest e-commerce player of European origin. With over 250 million offers, we’re now expanding our successful marketplace model across Central and Eastern Europe. Join us!

    Company profile

    Tech stack

      SQL

      advanced

      Team Player

      advanced

      Python

      regular

      Big Data

      nice to have

    Job description

    The salary range for this position (mid) is 12 300 - 17 600 PLN in gross terms (contract of employment)

    A hybrid work model that incorporates solutions developed by the leader and the team


    The Data Science Hub (DSH) is where we solve various business problems using analytical techniques and machine learning. We deliver insights and make decisions based on terabytes of data processed on a daily basis. Our team is a great place for people who seek continuous development opportunities and a unique chance to acquire interdisciplinary knowledge about how e-commerce platforms work. The variety of impacted business domains is best described by a diverse portfolio of projects, including:

    • logistics - delivery time prediction, logistic network optimization
    • marketing - category recommendation, next purchase prediction
    • pricing - price optimization
    • finance - sales forecasting
    • and many more…


    The Data Science Hub consists of 5 teams:

    • 3 Data Science teams
    • Data Analytics team
    • Data Engineering team


    And we are looking for new members for the Data Engineering team where we focus on the data processing and preparation, deployment and maintenance of our projects and sharing our skills with the rest of the team. 

    Join our team to enhance your skills related to deploying state-of-the-art data processing techniques and MLOps Machine Learning approaches.


    We are looking for people who:

    • Have ability to fluently work with SQL in traditional engines (e.g. MySQL, PostgreSQL) or cloud engines (e.g. BigQuery, Snowflake). You will be working with SQL on a daily basis.
    • Have experience in Python programming and are familiar with software engineering best practices (PEP8, clean architecture, code review, CI/CD etc.)
    • Have positive attitude and ability to work in a team
    • Are eager to constantly develop and broaden their knowledge


    An additional advantage would be:

    • Experience with Big Data ecosystem (Spark, Airflow)
    • Knowledge of BigData tools in Google Cloud Platform or other public cloud (e.g AWS, Azure)
    • Commercial experience in DevOps and CI/CD practice (e.g. GitHub Actions) in the area of ML/AI
    • Experience with cloud applications architecture


    Our techstack:

    • Python
    • Google Cloud Platform (AirFlow, BigQuery, Composer)
    • GitHub (code storage, CI/CD, hosting our own Data Science Python library)


    What we offer:

    • A hybrid work model that you will agree on with your leader and the team. We have well-located offices (with fully equipped kitchens and bicycle parking facilities) and excellent working tools (height-adjustable desks, interactive conference rooms)
    • Annual bonus up to 10% of the annual salary gross (depending on your annual assessment and the company's results)
    • A wide selection of fringe benefits in a cafeteria plan – you choose what you like (e.g. medical, sports or lunch packages, insurance, purchase vouchers)
    • English classes that we pay for related to the specific nature of your job
    • Working in a team you can always count on — we have on board top-class specialists and experts in their areas of expertise
    • A high degree of autonomy in terms of organizing your team’s work; we encourage you to develop continuously and try out new things
    • Hackathons, team tourism, training budget and an internal educational platform, MindUp (including training courses on work organization, means of communications, motivation to work and various technologies and subject-matter issues)


    What will your responsibilities be?

    • You will be actively responsible for building data processing tools for modeling and analysis – in close cooperation with both Data Science teams
    • You will be supporting both Data Science teams in the development of data sources for ad-hoc analyses and Machine Learning projects
    • You will process terabytes of data using Google Cloud Platform BigQuery, Composer, Dataflow and PySpark as well as optimize processes in terms of their performance and GCP cloud processing costs
    • You will collect process requirements from project groups and automate tasks related to preprocessing and data quality monitoring, prediction serving, as well as Machine Learning model monitoring and retraining
    • You will be responsible for the engineering quality of each project and you will cooperate with your colleagues on the engineering excellence


    Why is it worth working with us?

    • Through the supplied data and processes, you will have a meaningful impact on the operation of one of the largest e-commerce platforms in the world
    • Thanks to the wide range of projects we are involved in, you will never be without an interesting challenge to take on
    • You will have access to vast datasets (measured in petabytes)
    • You will get a chance to work in a team of experienced engineers and BigData specialists who are willing to share their knowledge (incl. with the general public, as part of allegro.tech)
    • Your professional growth will follow the most recent open-source technological trends
    • You will have an actual impact on the directions of product development and on the selection of particular technologies – we use the most recent and best technological solutions available, because we align them closely with our needs
    • We are a full-stack provider – we design, code, test, deploy and maintain our solutions


    Apply to Allegro and see why it is #dobrzetubyć (#goodtobehere)