We are looking for an experienced AWS Data Solutions Architect. A key focus will be building and managing streaming data pipelines and migrating data from RabbitMQ to Apache Pulsar.
Key Responsibilities:
- Design and develop streaming data solutions using Databricks, Spark, PySpark, and Pandas.
- Lead the migration of streaming data from RabbitMQ to Apache Pulsar.
- Develop and manage data models (conceptual, logical, physical) for large-scale data lakes and data warehouses.
- Document data architecture and create migration plans.
Requirements:
- Minimum of 4 years of experience in AWS data architecture.
- Experience with streaming data tools such as Databricks, Spark, PySpark.
- Proficiency in AWS technologies including Lambda, Kinesis, Glue, S3, Redshift.
- Experience migrating data from on-premise systems to the cloud (AWS).
- Familiarity with databases such as SQL Server, Oracle, PostgreSQL.
Preferred Skills:
- Knowledge of CI/CD tools like AWS CodePipeline, Terraform.
- Experience with containerization (Docker, Kubernetes).
- Familiarity with data modeling tools (ERwin, Power Designer).