We are looking for a skilled and motivated Hadoop Developer to join our client data engineering team in Warsaw. In this hybrid role, you’ll be responsible for designing, developing, and maintaining large-scale data pipelines in the AWS ecosystem, working with a variety of tools and technologies including Scala, Apache Spark, Kafka, and Cassandra. You’ll play a key role in building scalable and reliable data solutions that power analytics and business insights across the organization.
- Design, build, and maintain large-scale data pipelines using Hadoop, Apache Spark, and Kafka.
- Develop scalable data processing applications using Scala (preferred) or Java.
- Work within the AWS ecosystem (e.g., S3, EMR, Lambda, Glue, Redshift) to build and deploy cloud-native data solutions.
- Perform complex data analysis using strong SQL skills across MySQL, Cassandra, and other data stores.
- Optimize data workflows for performance, scalability, and fault tolerance.
- Collaborate with data scientists, analysts, and other engineering teams to support data-driven decision making.
- Ensure data quality, integrity, and governance across the platform.
- Troubleshoot, monitor, and improve performance of existing data pipelines.
- 4+ years of experience in big data engineering or a similar role.
- Proficiency in Scala (preferred) or Java for backend and data processing applications.
- Hands-on experience with Apache Spark, Hadoop, Kafka, MySQL, and Cassandra.
- Strong experience working with AWS services, especially for big data solutions.
- Advanced SQL skills for querying and transforming large datasets.
- Experience designing and optimizing data pipelines in production environments.
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills.