- responsible for at-scale infrastructure design, build and deployment with a focus on distributed systems,
- building and maintaining architecture patterns for data processing, workflow definitions, and system to system integrations using Big Data and Cloud technologies,
- evaluating and translating technical design to workable technical solutions/code and technical specifications at par with industry standards,
- driving creation of re-usable artifacts,
- establishing scalable, efficient, automated processes for data analysis, data model development, validation, and implementation,
- working closely with analysts/data scientists to understand impact to the downstream data models,
- writing efficient and well-organized software to ship products in an iterative, continual release environment,
- contributing and promoting good software engineering practices across the team,
- communicating clearly and effectively to technical and non-technical audiences,
- defining data retention policies,
- monitoring performance and advising any necessary infrastructure changes.
-
available to start immediately,
- readiness to work with the US Timezone (up to 9 PM CET),
- 5+ years’ experience as a data engineer,
- 2+ years' experience with AWS,
- Strong SQL skills,
- Python scripting proficiency,
- hands-on experience in building data processing pipelines,
- experience in structuring and modelling data in both relational and non-relational forms,
- strong expertise in cloud data platforms/warehouses like Databricks or Snowflake,
- extensive experience in big data engineering on a terabyte scale, including streaming technologies and near-real-time processing,
- experience working with VCS like Git,
- excellent command of oral and written English,
- ability to work with different stakeholders and drive consensus within the team
- good verbal and written communication skills in English,
- work from the European Union region and a work permit are required.
- proficiency in designing and implementing ETL/ELT processes and data integration workflows using tools like Apache Airflow, AWS Glue,
- understanding of big data and DevOps technologies (Kafka, Spark, Helm, Terraform),
- experience in CI/CD for the data environment,
- experience in testing for data processing,
- ML models operationalization (e.g., in Docker, Kubernetes).
CV review – HR call – Interview – Client Interview – Decision
🎁 Benefits 🎁
✍ Development:
- development budgets of up to 6,800 PLN,
- we fund certifications e.g.: AWS, Azure,
- access to Udemy, O'Reilly (formerly Safari Books Online) and more,
- events and technology conferences,
- technology Guilds,
- internal training,
- Xebia Upskill.
🩺 We take care of your health:
- private medical healthcare,
- multiSport card - we subsidise a MultiSport card,
- mental Health Support.
🤸♂️ We are flexible:
- B2B or employment contract,
- contract for an indefinite period.