For our business partner (trading company from USA) we are looking for Data Infrastructure Engineer!
Responsibilities on the project:
- Build and run data platform using such technologies as public cloud infrastructure (AWS and GCP), Kafka, databases and containers
- Develop data science platform based on open source software and Cloud services
- Build and run ETL tools and frameworks to onboard data into the platform, define schema, build DAG processing pipelines and monitor data quality
- Help develop machine learning development framework and pipelines
- Manage and run mission crucial production services
Requirements:
-
5+ years of experience in a Data Platform/Infrastructure Engineering role
- Strong experience with Python
- Experience building ETL and stream processing tools and frameworks using Kafka, Spark, Flink, Airflow/Prefect, etc.
- Experience with SQL and databases/engines such as MySQL, PostgreSQL, MS SQL, Snowflake, Redshift, Presto, etc.
- Familiarity with data science stack: e.g. Jupyter, Pandas, Scikit-learn, Dask, PyTorch, MLFlow, Kubeflow, etc.
- Experience with using AWS/GCP (S3/GCS, EC2/GCE, IAM, etc.), Kubernetes and Linux in production
- Strong proclivity for automation and DevOps practices
- Experience with managing increasing data volume, velocity and variety
-
English at least B2+/C1 level
-
University degree (preferred in IT or similar)
- Experience with: Java, C++, Rust, Go
- Understands TCP/IP and distributed systems
- Experience managing time series data
- Familiarity with working with open source communities
- Financial Services experience
-
Location: 100% remote
-
Wages: 160 PLN/H - 210 PLN/H net + VAT / B2B
-
Working hours: Overlap with US at least 5h (11:00/12:00 - 19:00/20:00 polish time)