I looking for Regular and Senior Data Engineers for Client from pharmaceutical industry.
Must have for Regular Engineer:
- 2+ years of working in Python or R
- 1+ years of experience with different types of storage (filesystem, relation, MPP, NoSQL) and working with various kinds of data (structured, unstructured, metrics, logs, etc.)
- 1+ years of experience in working in data architecture concepts (in any of following areas data modeling, metadata mng., workflow management, ETL/ELT, real-time streaming, data quality, distributed systems)
- Very good knowledge of code management tools (e.g. Git, SVN) and DevOps tools (e.g. Docker, Bamboo, Jenkins)
Must have for Senior Engineer:
- 4+ years of working in Python or R
- 3+ years of experience with different types of storage (filesystem, relation, MPP, NoSQL) and working with various kinds of data (structured, unstructured, metrics, logs, etc.)
- 3+ years of experience in working in data architecture concepts (in any of following areas data modeling, metadata mng., workflow management, ETL/ELT, real-time streaming, data quality, distributed systems)
- Very good knowledge of data serialization languages such as JSON, XML, YAML
- Excellent knowledge of code management tools (e.g. Git, SVN) and DevOps tools (e.g. Docker, Bamboo, Jenkins)
Nice to have for both profile:
- Experience working on GCP, AWS or other cloud platform
- Exposure to open source and proprietary cloud data pipeline tools such as Airflow, Glue and Dataflow
- Very good knowledge of relational databases