5656 2 x Mid+/Senior Flink Data Engineer
Remote from Poland
Budget: 165PLN/h + VAT
Our company is a global digital platform operating at massive scale, serving hundreds of millions of users worldwide. We focus on building consumer-facing subscription products that connect users with content creators through personalized, data-driven experiences. Our culture emphasizes collaboration, experimentation, and engineering excellence in a highly distributed, cloud-native environment.
Project Overview
The project centers on the continuous development and modernization of a large-scale streaming data platform. The main objective is to evolve existing Apache Flink-based streaming pipelines while executing complex data migrations and platform upgrades. Key responsibilities include:
Developing and enhancing real-time streaming pipelines using Apache Flink
Migrating existing Flink jobs using the DataStream API and adapting them to newer platform standards
Leading and executing the upgrade of the Flink platform to version 2.0
Designing, optimizing, and maintaining high-throughput, fault-tolerant streaming architectures
Migrating large-scale datasets from BigQuery (BQ) to Data Cloud Storage (DCS)
Scaling and automating ongoing data migration processes to support growing data volumes
Converting datasets from Avro to Parquet format with emphasis on performance, schema evolution, and storage optimization
Leveraging AI-powered tools to accelerate migration, validation, and transformation workflows
Ensuring data quality, integrity, and minimal downtime during migrations
Collaborating with cross-functional teams and effectively communicating technical concepts to non-technical stakeholders
Key Requirements
Strong hands-on experience with Apache Flink, especially development using the DataStream API
Proven experience maintaining and upgrading Flink environments, ideally with exposure to Flink 2.0
Deep understanding of streaming pipeline architecture, performance tuning, state management, and fault tolerance
Experience migrating large-scale datasets from BigQuery (BQ) to Data Cloud Storage (DCS)
Strong proficiency in data format conversion, particularly Avro to Parquet
Ability to design, scale, and automate migration workflows while ensuring data integrity and minimal service disruption
Solid knowledge of Google Cloud Platform (GCP) and its data services
Good understanding of distributed systems, schema evolution, and storage optimization strategies
Ability to break down complex migration and platform challenges into clear, actionable steps
Proactive mindset with strong ownership of solutions and risk identification
Clear and effective communication skills, especially in explaining technical topics to non-technical stakeholders
Nice to Have
Interest in and familiarity with emerging AI-driven practices, with a willingness to explore and experiment beyond standard approaches.
Experience working on high-scale, consumer-facing data platforms
Background in long-running migration programs involving multiple data sources and formats
Familiarity with observability, monitoring, and alerting for streaming systems
Practical experience using AI-powered assistants to improve productivity, quality, or decision-making in software delivery.
5656 2 x Mid+/Senior Flink Data Engineer
5656 2 x Mid+/Senior Flink Data Engineer