return to list

๐Ÿค“ Data Engineer

14 000 - 17 000 PLN gross

๐ŸŒ Dotcommunity | Armii Krajowej, Krakรณw

๐Ÿ–ฅ http://www.dotcommunity.eu/pl-welcome.html

๐Ÿ‘‰ Requirements:

  • SQL ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ
  • data processing &nbsp ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ
  • data storage &nbsp ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ
  • English ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ
  • Big Data ๐Ÿ”ธ๐Ÿ”ธ๐Ÿ”ธ
  • ๐Ÿ‘‰ Nice to have:

  • Spark ๐Ÿ”ธ
  • AWS ๐Ÿ”ธ
  • yarn &nbsp ๐Ÿ”ธ
  • airflow &nbsp ๐Ÿ”ธ
  • Microservices ๐Ÿ”ธ
  • We are looking for a candidate to join our client's team as a Data Engineer. They are a software house which operates in the advertising and media industry, located at Armii Krajowej street.

    Advertising Solutions is a relatively new area in Media organisation which houses engineering teams for our back-office systems used by various sales organisations at the company. These systems include Rose which is used to book advertising campaigns and Vantage which provides campaign reporting.


    ABOUT THE TEAM

    They are now looking to establish a team to own and operate the Advertising API that underpins these products, along with others within Schibsted. You will be off to a running start and will have to learn the ropes of existing systems with the help of their established teams, plan and execute on the hand-over from the current team in London. Expect initial travels to London and/or hosting the London team members locally.

    Once you learn the system you and your teammates will continuously work on its technical evolution, scaling and simplification. You will be expected to be an active participant when deciding how to implement new features together with the neighbouring teams that depend on you for their work.


    SKILLS & REQUIREMENTS

    They handle more than 250000 campaigns, 100000 advertisers, and more than 140 publishers across 20 different countries. About 1.5 TB of data is processed every day using more than 100 Spark jobs.

    Their data pipeline is built on top of AWS EMR, Spark, Yarn, Airflow and microservices based on Twitter Finatra framework. Apache Avro and Parquet are used for data serialization and schema definition/evolution.

    They donโ€™t expect you to have experience with all the technologies that they use but it would be good if you know at least some of them or have worked with similar ones.