Head of Data Engineering

Data

Head of Data Engineering

Data
Powstańców Warszawy 6, Sopot

Autopay S.A.

Full-time
B2B, Permanent
Senior
Hybrid

Job description

About the company

Autopay Global is the newest member of the Autopay family, aiming to expand the reach of the group’s state-of-the-art payment integration and payment data technologies to the international market, providing seamless integration with local PSPs, support for multiple currencies and compliance with local frameworks. We have a very forward-looking approach to our products, we value creativity, passion and drive to leverage the newest achievements in technology to our advantage.

 

To support our dynamic expansion, we are looking for a new Head of Data Engineering for a full-time, hybrid work in Warsaw or Gdańsk.

About the role

The Head of Data Engineering owns the end-to-end data architecture and execution, with hands-on depth in PySpark and Databricks and strong experience building AI-ready data foundations on Google Cloud Storage (GCS) and Google Vertex AI.

You will be responsible for delivering a secure, scalable, and low-latency data lakehouse and feature platform that enables Autopay’s AI core (agents, RAG, ranking, decisioning) and activation systems to run on reliable, high-quality, well-governed data across batch and streaming. You will also be responsible for hiring, leading and mentoring a team of high-performing data engineering proffessionals.   

 

  •       Define the lakehouse reference architecture on GCS with Databricks/Delta Lake, 
  •       build and operate PySpark pipelines in Databricks for both streaming and batch workloads,
  •       implement streaming ingestion,
  •       own the Customer 360 / CDP layer: unify events, transactions, and user identifiers, 
  •       deliver a real-time feature layer (feature store) that publishes segments, scores, and vectors, 
  •       create and maintain embeddings and retrieval indexes to power RAG in Autopay AI Core (chunking strategies, metadata, refresh policies, and retrieval evaluation, 
  •       establish data governance with Dataplex/Data Catalog and/or Unity Catalog, 
  •       own data observability for pipelines: freshness, completeness, schema drift, anomaly detection, and automated remediation workflows.

What tools will you be working with?

  •        Technology: PySpark, Databricks, Google Cloud Storage, Google Vertex AI, Delta Lake
  •        Nice to have: Experience with identity resolution inputs, experience building near-real-time segmentation, CLV, and propensity scoring pipelines, familiarity with vector databases and multi-cloud data movement patterns. 

Requirements and skills we are looking for in a person hired for this role:

  •       10+ years in data engineering and still hands-on to build ground up forming a team; 3-5+ years leading data platform teams with ownership of production data SLAs,
  •       deep hands-on expertise with PySpark and Spark performance tuning (shuffle optimization, partitioning, checkpointing, incremental loads),
  •       strong experience with Databricks (jobs/workflows, Delta Lake, governance) and building lakehouse architectures on GCS,
  •       proven delivery of streaming + batch data platforms that power real-time product experiences (not just analytics),
  •       experience building feature stores and ML-ready datasets with point-in-time correctness and strong governance,
  •       strong grasp of privacy and compliance in data systems: PII handling, consent, and auditability,
  •       Google Vertex AI experience: building data pipelines that feed training, evaluation, and inference workflows; understanding of dataset/version management, 
  •       hands-on experience supporting RAG systems: document ingestion, chunking, embedding generation, retrieval evaluation, and index refresh strategies,
  •       experience with retrieval-aware training approaches (e.g., retrieval augmented fine-tuning / RAFT) and producing high-quality supervised datasets with provenance,
  •       ability to collaborate with AI Engineers on MCP-based tools and agent workflows (tool schemas, rate limits, caching, and audit logs).

What we offer

  •       a lidership role in fast-growing, global fintech company,
  •       possibility to work with cutting-edge tools and technologies,
  •       independence in decision-making, 
  •       friendly working environment, team support, no dress code. 

 

Join us and let's head together where no one has gone before!

 

Tech stack

    PySpark

    advanced

    Databricks

    advanced

    AI

    advanced

    Google Cloud

    regular

    Spark

    regular

Office location

Head of Data Engineering

Summary of the offer

Head of Data Engineering

Powstańców Warszawy 6, Sopot
Autopay S.A.
By applying, I consent to the processing of my personal data for the purpose of conducting the recruitment process. Informujemy, że administratorem danych jest Autopay S.A z siedzibą w Sopocie, ul. Powstańców Warszawy 6 (dalej jako "administrator"). ... MoreThis site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.