Data Engineer
At DAC.digital, we are constantly growing our business. As part of our growth strategy, we are strengthening our strategic partnership with a company specializing in crafting data models, dashboards, and data activation.
Key information:
Salary:
28 000 - 31 000 PLN net/month - pure B2B contract
24 500 – 27 000 PLN net/month – B2B contract (days off included)
It is vital that you have:
4+ years hands-on data engineering / platform experience;
strong experience with Spark/PySpark, Kafka or other streaming platforms, cloud-native data lakes or lakehouses (Iceberg, Delta Lake, Hudi, with Iceberg as a plus), dbt or similar transformation frameworks, and advanced SQL including performance tuning;
solid Python skills for data pipelines and integration work;
familiarity with NLP / LLM-driven text processing workflows (embedding models, classification, enrichment);
experience with BI / query engines such as Dremio, Trino, Presto, or BigQuery plus a viz tool (Metabase, Looker, Power BI, etc.);
experience deploying data platforms in cloud and/or on-prem environments - Docker, Kubernetes (or equivalent orchestration), CI/CD for data pipelines;
understanding of security, networking basics and how to operate inside restricted enterprise environments (no direct internet access, firewall rules, etc.);
knowledge of English (min. B2);
high communication skills;
eager to learn and share knowledge.
Nice to have:
experience with marketing / growth analytics, CDP, or telco data;
background in ML/AI beyond pipelines (e.g. recommendation systems, uplift modelling);
prior experience in consulting / client-facing roles;
some front-end exposure (Metabase customisation, simple React dashboards) if we want one person to bridge.
Technology stack:
Python
SQL
Apache Spark / PySpark
Kafka
Iceberg
Nessie
MinIO
Kubernetes
Helm Charts
You will be responsible for supporting our team in:
designing, implementing, and maintaining streaming and batch data pipelines using Kafka, Spark/PySpark, Apache Iceberg, and dbt;
managing schema evolution, partitioning, performance, and data quality in the Iceberg lakehouse;
building and maintaining NLP and enrichment flows;
integrating with LangChain and LangGraph-based components to resolve senders to companies, enrich data from public internet sources, and classify companies into brand categories and verticals such as F&B, footwear, and online brokers;
supporting experimentation with ML and NLP components, including monitoring and logging;
working with Dremio and Metabase to ensure performant queries and dashboards and expose data in a usable ways;
handling deployment and operations in constrained environments, including on-prem or behind-firewall deployments at MNOs, containerisation with Docker and Kubernetes, and logging and lightweight monitoring for services and pipelines;
ensuring stability, observability, and reproducibility of the platform.
What do we offer:
possibility to work 100% remotely or on-site at our office in Gdańsk;
b2b contract with included 30 days off;
private Medical care;
group insurance;
pre-paid card or Sport Card;
referral program;
real Agile practices;
employee well-being online platform;
cafeteria benefits.
Data Engineer
Data Engineer