#1 Job Board for tech industry in Europe

Senior Data Engineer (w/ DevOps)
Data

Senior Data Engineer (w/ DevOps)

Poznań
Type of work
Full-time
Experience
Senior
Employment Type
B2B
Operating mode
Remote

Tech stack

    Airflow

    advanced

    Data

    advanced

    Terraform

    advanced

    Amazon AWS

    advanced

    Python

    advanced

    Polish/English

    advanced

    Snowflake

    regular

    SQL

    regular

Job description

Online interview

Advance your career with Sunscrapers, a leading force in software development, now expanding its presence in a data-centric environment. Join us in our mission to help clients grow and innovate through a comprehensive tech stack and robust data-related projects. Enjoy a competitive compensation package that reflects your skills and expertise while working in a company that values ambition, technical excellence, trust-based partnerships, and actively supports contributions to R&D initiatives.


The project:

We are carrying out the project for our client, an American private equity and investment management fund - listed on the Forbes 500 list - based in New York.

We support them in the area of the infrastructure and data platform, and very recently we also build and experiment with Gen AI applications. The client operates very widely in the world of finance, loans, investments and real estate.


As a Senior Data Engineer you’ll design and implement core systems that enable data science and data visualization at companies that implement data-driven decision processes to create a competitive advantage. 

You’ll build data platform for data and business teams, including internal tooling, data pipeline orchestrator, data warehouses and more, using:


Technologies: Python, Terraform, SQL, Pandas, Shell scripts

Tools: Apache Airflow / Astronomer, GIT, Docker, Kubernetes, Snowflake, Pinecone, Neo4j, Jenkins, Jupyter Notebook, OpenAI API, Artifactory, Windows with WSL, Linux, Gitlab

AWS: EC2, ELB, IAM, RDS, Route53, S3, and more

Best Practices: Continuous Integration, Code Reviews


The ideal candidate will be well organized, eager to constantly improve and learn, driven and, most of all - a team player!


Your responsibilities will include:

  • Developing PoCs using latest technologies, experimenting with third party integrations
  • Delivering production grade applications once PoCs are validated
  • Creating solutions that enable data scientists and business analysts to be self-sufficient as much as possible.
  • Finding new ways how to leverage Gen AI applications and underlying vector and graph data storages
  • Designing datasets and schemes for consistency and easy access
  • Contributing data technology stacks including data warehouses and ETL pipelines
  • Building data flows for fetching, aggregation and data modeling using batch and streaming pipelines
  • Documenting design decisions before implementation


Requirements


What's important for us?

  • At least 7+ years of professional experience in data-related role
  • Experience with infrastructure-as-code tools, like Terraform
  • Proficiency in using Airflow
  • Expertise in AWS stack and services (EC2, ELB, IAM, RDS, Route53, S3, and more)
  • Proficiency in using Docker
  • Expertise in Python and SQL languages
  • Experience with data warehouses (Snowflake)
  • Experience with different types of database technologies (RDBMS, vector, graphs, document based, etc.)
  • Great analytical skills and attention to detail - asking questions and proactively searching for answers
  • Excellent command in spoken and written English, at least C1
  • Creative problem-solving skills
  • Excellent technical documentation and writing skills
  • Ability to work with both Windows and Unix-like operating systems as the primary work environments
  • Undergraduate or graduate degree in Computer Science, Engineering, Mathematics, or similar


You will score extra points for:

  • Experience with integrating LLMs (OpenAI but also others, maybe open source)
  • Familiarity with data visualization in Python using either Matplotlib, Seaborn or Bokeh
  • Proficiency in statistics and machine learning, as well as Python libraries like Pandas, NumPy, matplotlib, seaborn, scikit-learn, etc
  • Knowledge of any Python web framework, like Django or Flask with SQLAlchemy
  • Experience in operating within a secure networking environment, like a corporate proxy
  • Experience in working with repository manager, for example Jfrog Artifactory



Benefits


What do we offer?

  • Working alongside a talented team of software engineers who are changing the image of Poland abroad
  • Culture of teamwork, professional development and knowledge sharing (https://www.youtube.com/user/sunscraperscom)
  • Flexible working hours and remote work possibility
  • Comfortable office in central Warsaw, equipped with all the necessary tools for conquering the universe (Macbook Pro/Dell, external screen, ergonomic chairs)

Sounds like a perfect place for you? Don’t hesitate to click apply and submit your application today!

Check similar offers

Principal BI Engineer

New
Cornerstone OnDemand
3.62K - 5.79K USD
Gdańsk
, Fully remote
Fully remote
Tableau
SQL
Snowflake

Senior Data Analyst

New
EER POLAND
6.5K - 7.5K USD
Poznań
, Fully remote
Fully remote
SQL
B2C Analytics
User Behavior Analytics

Senior Software Engineer

New
Ciklum
5.15K - 5.67K USD
Gdańsk
, Fully remote
Fully remote
JavaScript
Azure Services
Powershell

Mid/Senior Data Engineer with GCP

New
Holisticon Insight
5.55K - 6.47K USD
Warszawa
, Fully remote
Fully remote
English
GCP
PySpark

Azure DataBricks Engineer

New
CRODU
5.67K - 7.7K USD
Szczecin
, Fully remote
Fully remote
Python
NoSQL
Azure