In this role you will join a project where we are building an ecosystem around an Enterprise Data Lake. Our solution is mainly built in Scala language and we help developers by providing automation and pipelines for testing and deployment. We work in an agile setup in an international environment.
Responsibilities:
- Developing Scala/Spark programs, scripts, and macros for data extraction, transformation and analysis
- Designing and implementing solutions to meet business requirements
- Supporting and maintaining existing Hadoop applications and related technologies
- Developing and maintaining technical documentation, including data models, process flows and system diagrams
Requirements:
- Min. 5 years of experience in Scala/Spark development
- Creating Scala/Spark jobs for data transformation and aggregation as per the complex business requirements
- Ability to work in a challenging and agile environment with quick turnaround times and strict deadlines
- Experience in performing Unit tests of the Scala code
- Raise PR, trigger build and release JAR versions for deployment via Jenkins pipeline
- Familiarity with CI/CD concepts and the processes
- Experience in peer review the code
- Performing RCA of the bugs raised
- Excellent understanding of Hadoop ecosystem
- Well versed with the majority of below technologies: SQL/HQL (Hive Queries), Oozie, Jenkins, ETL, Shell scripting, GIT, Splunk
Our Offer:
-
100% remote
- MultiSport Plus
- Group insurance
- Medicover Premium
- e-learning platform