Together with our partner
ToolsGroup, we want to invite a Data Engineer for a cooperation in a long-term project.
ToolsGroup is the world’s most experienced provider of market-driven demand analytics, supply chain optimization, and retail demand management software. Having been high performing for more than 20 years, they have a loyal global customer base of more than 300 customers worldwide in more than 44 countries and an excellent reputation for delivering high-quality products and superior results. ToolsGroup is recognized in multiple Gartner Magic Quadrants and Nucleus Value Matrices.
About the role:
As a Data Engineer, you will be responsible for assisting our Data Scientists and Machine Learning engineers in designing and executing experiments and in industrializing advanced AI & ML features. You will also be involved in helping software engineers in implementing data structures and workflows for real-time analytics at scale.
What you will take on:
- Designing reliable and scalable ETL/ELT pipelines right from ingestion of the data to the delivery of the end product;
- Building custom solutions to automate workflows and supporting architecture;
- Understanding business logic, allowing you and other team members to make informed decisions about design changes and improvements;
- Taking ownership of existing products and pipelines, making sure they are delivered on time;
- Implementing upgrades and optimizations to existing processes, including migration of the code to newer versions of libraries/packages/infrastructure/etc.;
- Communicating with different stakeholders to set expectations and figure out priorities;
- Ensuring quality, reliability and uptime for critical automated processes, including helping the data science team diagnose and resolve issues in the pipeline and in the data;
- Designing environments within the ToolsGroup platform to house the data in a secure manner and enabling clients and internal teams to access it appropriately.
What you already have:
- At least 3 years of commercial experience programming in languages typically employed in AI/ML such as Python, Java, Scala, GoLang, or R;
- At least 2 years of experience working with distributed systems such as Hadoop, Spark in cloud computing platforms such as AWS, GCP, and Azure (AWS Glue, Databrick, Azure Data Factory);
- At least 1 year of experience maintaining a client-facing data pipeline in a production environment;
- Strong experience working with Apache Kafka;
- Experience in working in an Agile environment at a software company is a plus;
- Great communication and teamwork skills;
- Experience with both structured and unstructured databases, SQL, No-SQL, Relational Databases, Datalakes, and Lakehouses. Bonus points for having experience with Graph Databases and Knowledge of Graphs;
- Knowledge of engineering best practices.
Preferred skills:
- Good understanding of Data Engineering, NoSQL databases and database design, distributed systems, and/or information retrieval;
- Knowledge of Apache Airflow;
- Familiarity with common Azure services;
- Experience writing high-performance queries in SQL and unstructured or graph query languages;
- Experience with SaaS highly distributed systems, distributed cloud storage, event queues, and data streaming using Apache Kafka.
- Ability to plan and collect requirements for projects, and interact with the analyst and data science teams.
In exchange for your skills we offer:
- Supportive relationships, built on transparency and a flat structure, in a diverse and multinational team;
- Office in the center of historical Kraków, where your dog is always welcome;
- Flexibility when it comes to working from the office or home;
- Perks incl Multikafeteria, Group Life Generali insurance, Luxmed, Multisport, language lessons 1:1;
- Sport and other events, including, weekly running, squash, and team lunches on the house;
- Free tea, coffee, and all-you-can-eat fruits and nuts in the office.
Educational Requirements:
- Undergraduate degree in Computer Science, Computer Information Systems, Information Systems, or other related fields.
The recruitment process includes:
- HR screening 30min zoom call;
- 1h technical interview;
- General call with the team leader;
- Making an offer meeting.