#1 Job Board for tech industry in Europe

Senior Backend Engineer (with Data Science Skills) at PeakData
New
Data

Senior Backend Engineer (with Data Science Skills) at PeakData

7 753 - 9 969 USD/monthNet per month - B2B
7 753 - 9 969 USD/monthNet per month - B2B
Type of work
Full-time
Experience
Senior
Employment Type
B2B
Operating mode
Hybrid

Tech stack

    English

    B2

    AWS

    advanced

    Communication Skills

    advanced

    Docker

    advanced

    English

    advanced

    SQL

    advanced

    Python

    advanced

    LLM

    advanced

    GCP

    regular

Job description

Online interview

About PeakData

PeakData provides AI-powered market intelligence to optimize drug launch execution and resource allocation for pharmaceutical companies. Our platform delivers actionable insights on healthcare professionals (HCPs) and healthcare organizations (HCOs), empowering commercial leaders with real-time, data-driven decision-making.



Role Overview

We're looking for a Senior Engineer with strong data science capabilities to join our Data Platform team. In this role, you’ll design and build cloud-native data solutions that support large-scale processing, analytics, and AI-powered automation across our platform.

This is a hands-on, senior-level role. You will be expected to work independently, own end-to-end pipelines and infrastructure, and drive initiatives forward both individually and within the team. You should have a strong foundation in Python, SQL, AWS, and/or GCP, with experience using or integrating LLMs into data workflows.



Tech Environment

You’ll work with and expand upon:

  • Python for data pipelines and automation

  • SQL (PostgreSQL) for transformation and analytics

  • AWS (S3, Glue, Lambda, ECS, Bedrock) as primary cloud environment

  • GCP (Vertex AI) for select workloads and integrations

  • Medallion architecture with RAW/CLEANED/CURATED layers

  • LLM integrations for automation, enrichment, and insight generation

  • Data quality frameworks and orchestration tools (e.g., Argo)



Key Responsibilities

Engineering Ownership

  • Design, implement, and maintain scalable and efficient data pipelines across AWS and GCP

  • Build data products and services supporting both internal analytics and client-facing insights

  • Own ETL/ELT workflows from ingestion to curation

  • Implement observability and alerting for pipeline health and data integrity

  • Integrate LLMs into workflows to support enrichment, automation, or intelligent data handling


Team Leadership & Initiative

  • Act as a technical lead for data engineering projects, driving execution independently

  • Collaborate cross-functionally with Data Science, Product, and Engineering teams

  • Contribute to architectural decisions and long-term data platform evolution

  • Champion best practices for performance, security, and scalability


Data Science & LLM Integration

  • Apply data science techniques where appropriate (e.g., clustering, statistical inference)

  • Prototype and validate LLM-powered solutions using tools like AWS Bedrock or Vertex AI

  • Use prompt engineering and evaluation frameworks to fine-tune LLM interactions

  • Help bridge engineering and AI innovation across the platform



Qualifications

Required Skills & Experience

  • 6+ years of experience in data engineering or back-end systems with data-heavy workloads

  • Strong hands-on skills with Python and SQL

  • Deep understanding of AWS cloud data tooling (S3, Lambda, Glue, Step Functions, etc.)

  • Working experience with GCP services, especially BigQuery and Vertex AI

  • Exposure to LLMs and how they integrate into data workflows

  • Experience building data pipelines at scale with monitoring and alerting

  • Ability to work independently and take ownership of technical topics


Bonus Skills

  • Experience with Argo, Airflow or similar orchestration frameworks

  • Familiarity with IaC tools (Terraform) for deploying infrastructure

  • Experience with data quality monitoring, validation frameworks, or anomaly detection

  • Previous work in healthcare, life sciences, or regulated data environments



Personal Attributes

  • Regular visits to our Wrocław office (ideally twice per month) are very welcome and positively received — we value face-to-face collaboration when possible!

  • Proactive: You take initiative and don’t wait for tasks to be assigned

  • Autonomous: You can own projects from design to production with minimal oversight

  • Curious: You explore new approaches (especially LLMs/AI) and bring them to the table

  • Collaborative: You work well with cross-functional teams

  • Customer-aware: You understand the real-world impact of your pipelines and models



What We Offer

  • Purpose-driven work: support pharmaceutical innovation and better patient outcomes

  • Ownership: real autonomy in shaping our data systems and how they scale

  • Innovation: work on LLM integration and next-gen data workflows

  • A collaborative, fast-moving environment

  • Competitive compensation

  • Access to both AWS and GCP ecosystems in production



If you're a hands-on data engineer who enjoys owning end-to-end systems, loves solving real business problems, and thrives in a hybrid cloud + AI environment — we want to talk to you.

7 753 - 9 969 USD/month

Net per month - B2B