#1 Job Board for tech industry in Europe

Data Engineer
New
Data

Data Engineer

4 845 - 5 853 USD/monthNet per month - B2B
4 845 - 5 853 USD/monthNet per month - B2B
Type of work
Full-time
Experience
Mid
Employment Type
B2B
Operating mode
Remote

Tech stack

    English

    B2

    Data

    master

    AWS

    advanced

    Azure

    advanced

    GCP

    advanced

    Databricks

    regular

Job description

We are looking for an experienced Data Engineer responsible for planning, developing, and maintaining cloud environments for our clients. 

About Devapo

At Devapo, we focus on continuous self-development and acquiring new knowledge. If you are a fast learner, want to participate in international projects, are a team player, and can work independently — join us!

We provide our clients with more than just code — we want to equip them with tools that allow their businesses to flourish. Our clients’ success is our success, which is why we ensure that everyone who creates Devapo has a long-term goal in mind.

Key Responsibilities:

Design, implement, and maintain scalable and efficient data pipelines in one of the cloud environments (Azure, AWS, GCP) using tools such as Databricks, Glue, Dataflow, or Azure Data Factory

  • Develop and optimize ETL/ELT processes using cloud-native services (e.g., Azure Data Factory, AWS Glue, GCP Dataflow) and Apache Spark/Databricks

  • Build Big Data solutions aligned with business and analytical requirements across cloud platforms

  • Collaborate with Data Science, BI, and development teams to deliver high-quality, well-structured, and performant data

  • Monitor and improve the performance, reliability, and scalability of data processing systems

  • Implement robust data governance, security standards, and best practices across cloud environments

  • Research and evaluate new tools and technologies within the cloud and data engineering ecosystem

Requirements:

  • Minimum 3 years of experience as a Data Engineer or in a similar role

  • Hands-on experience with one or more major cloud platforms (Azure, AWS, GCP); deep knowledge of cloud data services such as:

  • Azure Data Factory, Azure Data Lake, Synapse Analytics (Azure)

  • AWS Glue, S3, Redshift, Athena (AWS)

  • GCP Dataflow, BigQuery, Cloud Storage (GCP)

  • Extensive experience with Databricks and Apache Spark

  • Proficiency in SQL and experience with relational and columnar databases

  • Strong programming skills in Python and PySpark

  • Experience designing and optimizing data pipelines in distributed, cloud-based architectures

  • Familiarity with Delta Lake or other modern data lake architectures

  • Solid understanding of data modeling and schema design

What We Offer:

  • Salary: 17 800 - 21 500 PLN (B2B contract)

  • Co-financing for training and certifications, as well as guaranteed time for learning during working hours

  • Private medical care and a Multisport card

  • Language classes (English)

  • Flexible working hours and the possibility of hybrid work (Warsaw)

  • Team integration meetings and company events

  • Employee referral program with a bonus

  • An individually tailored career development path

4 845 - 5 853 USD/month

Net per month - B2B