Data Architect (GCP, Snowflake)
Addepto is a leading AI consulting (https://addepto.com/ai-consulting/) and data engineering (https://addepto.com/data-engineering-services/) company that builds scalable, ROI-focused AI solutions for some of the world's largest enterprises and pioneering startups, including Rolls Royce, Continental, Porsche, ABB, and WGU. With an exclusive focus on Artificial Intelligence and Big Data, Addepto helps organizations unlock the full potential of their data through systems designed for measurable business impact and long-term growth.
The company's work extends beyond client engagements. Drawing from real-world challenges and insights, Addepto has developed its own product - ContextClue - and actively contributes open-source solutions to the AI community. This commitment to transforming practical experience into scalable innovation has earned Addepto recognition by Forbes as one of the top 10 AI consulting companies worldwide.
As part of KMS Technology, a US-based global technology group, Addepto combines deep AI specialization with enterprise-scale delivery capabilities—enabling the partnership to move clients from AI experimentation to production impact, securely and at scale.
As a Data Architect, you will lead data architecture and engineering initiatives within GCP environments, including designing, evolving, and governing data platforms built on GCP and Snowflake. You will work closely with clients to define data strategies, design scalable architectures, and guide implementation teams. You will play a key role in shaping future-state data architectures, roadmaps, and implementation plans, while ensuring data platforms are secure, scalable, and aligned with business goals.
📍 Location: This role requires on-site work in the United States
🚀 Your main responsibilities:
Lead data architecture and engineering projects within GCP and Snowflake environments, including migration, transformation, and platform design.
Design end-to-end data architectures (ingestion, storage, processing, and serving layers).
Define and guide implementation of data integration, ETL/ELT strategies, and pipeline architectures.
Develop logical and physical data models, schemas, and data structures across business domains.
Guide clients on data strategy, architecture decisions, and best practices across GCP.
Design and enforce data governance, security, and access control frameworks.
Work closely with security teams to ensure data protection and compliance.
Lead client workshops to identify data sources, flows, and requirements.
Define future-state architectures, roadmaps, and implementation plans.
Collaborate with cross-functional teams (data scientists, engineers, business stakeholders) to deliver data solutions.
Evaluate and select tools, frameworks, and architectural patterns.
Provide technical leadership and mentorship to engineering teams.
Oversee performance optimization, scalability, and cost efficiency of data platforms.
🎯 What you'll need to succeed in this role:
7+ years of experience in Data Engineering, Data Architecture, or Data Infrastructure roles.
3+ years of experience working with GCP or similar cloud platforms (AWS, Azure).
Proven experience in designing data architectures and large-scale data platforms.
Strong experience with GCP managed services (BigQuery, Cloud SQL, Cloud Spanner, Cloud Bigtable).
Hands-on experience with Snowflake, including data modeling and performance optimization.
Strong expertise in data modeling, database design, and data architecture patterns.
Experience with ETL/ELT design, data integration, and migration from legacy systems.
Experience working with structured, semi-structured, and unstructured data.
Understanding of data governance, security, and compliance frameworks.
Experience with Infrastructure as Code (Terraform, Ansible, or similar).
Proficiency in SQL and good understanding of Python.
Strong analytical, problem-solving, and communication skills.
Experience working in a client-facing or consulting environment.
➕ Nice to have:
Experience with data platform migrations to Snowflake or cloud environments.
Familiarity with data governance program implementation.
Knowledge of Kimball, Data Vault, or other data modeling methodologies.
Experience with multi-cloud or hybrid architectures.
Familiarity with data observability and monitoring tools.
Understanding of FinOps and cloud cost optimization.
🎁 Discover our perks & benefits:
Work in a supportive team of passionate enthusiasts of AI & Big Data.
Engage with top-tier global enterprises and cutting-edge startups on international projects.
Enjoy flexible work arrangements, allowing you to work remotely or from modern offices and coworking spaces.
Accelerate your professional growth through career paths, knowledge-sharing initiatives, language classes, and sponsored training or conferences, including a partnership with Databricks, which offers industry-leading training materials and certifications.
Choose your preferred form of cooperation: B2B or a contract of mandate, and make use of 20 fully paid days off.
Participate in team-building events and utilize the integration budget.
Celebrate work anniversaries, birthdays, and milestones.
Access medical and sports packages, eye care, and well-being support services, including psychotherapy and coaching.
Get full work equipment for optimal productivity, including a laptop and other necessary devices.
With our backing, you can boost your personal brand by speaking at conferences, writing for our blog, or participating in meetups.
Experience a smooth onboarding with a dedicated buddy, and start your journey in our friendly, supportive, and autonomous culture.
Are you interested in Addepto and would like to join us?
Get in touch! We are looking forward to receiving your application. Would you like to know more about us?
Visit our website (career page) and social media (Facebook, LinkedIn, Instagram).
Data Architect (GCP, Snowflake)
Data Architect (GCP, Snowflake)