#1 Job Board for tech industry in Europe

  • Job offers
  • Senior Spark Engineer
    Data

    Senior Spark Engineer

    Łódź
    Type of work
    Full-time
    Experience
    Senior
    Employment Type
    B2B
    Operating mode
    Remote
    Link Group

    Link Group

    Hundreds of IT opportunities are waiting for you—let’s make it happen! Since 2016, our team of tech enthusiasts has been building exceptional IT teams for Fortune 500 companies and startups worldwide. Join impactful projects in BFSI, CPG, Industrial, and Life Sciences & Healthcare industries. Work with cutting-edge technologies like Cloud, Business Intelligence, Data, and SAP. Unlock your potential, grow your skills, and collaborate with top global clients. Ready for your next big career move? Let’s link with us!

    Company profile

    Tech stack

      Spark

      advanced

      Cloud

      regular

      SQL

      regular

      Big Data

      regular

    Job description

    Online interview
    Friendly offer

    Employment Type: Full-Time, Remote

    Job Description: We are looking for a highly experienced Senior Spark Engineer with deep expertise in Apache Spark, particularly in performance tuning and managing cyclic Spark data flows. The candidate should be proficient in troubleshooting and optimizing real-time data processing systems, including customizing the Catalyst Optimizer. Experience working with federated data systems and distributed computing environments is essential, along with the ability to integrate external systems and APIs. The role will involve optimizing Spark pipeline performance across large-scale, multi-cloud environments.


    Key Responsibilities:


    • Expertise in Apache Spark: Utilize in-depth knowledge of Spark, including performance tuning, query optimization, and customizing the Catalyst Optimizer for distributed systems.
    • Federated Data Systems: Design, implement, and manage data workflows within federated models across multi-cloud environments.
    • Performance Optimization: Diagnose and address bottlenecks in Spark jobs, ensuring scalable and efficient performance on large clusters.
    • Distributed Computing: Manage Spark clusters, oversee task scheduling, resource allocation, and ensure fault tolerance in distributed environments.
    • API Integration: Connect Spark applications with external systems and APIs to improve data processing workflows.
    • Scala and Java Development: Apply strong skills in Scala and Java to build, maintain, and optimize real-time distributed applications in Spark.
    • Front-End Collaboration: Work with front-end developers and data teams to create and deploy user interfaces for monitoring Spark pipeline performance.
    • CI/CD and Version Control: Develop and manage CI/CD pipelines to ensure reliable software development practices, version control, and automated deployment for distributed applications.


    Required Skills and Experience:


    • Apache Spark: Advanced experience in tuning, optimizing, and customizing the Spark Catalyst Optimizer for maximum performance.
    • Scala and Java Proficiency: Strong hands-on experience with Scala and Java in Spark-based distributed systems.
    • Federated Data Models: Proven experience managing federated data systems in multi-cloud environments (e.g., AWS, GCP, Azure).
    • Distributed Computing: Deep understanding of distributed computing principles, including task scheduling, resource management, fault tolerance, and cluster optimization.
    • Performance Optimization: Demonstrated expertise in optimizing Spark pipelines for large-scale, high-volume systems.
    • API Integration: Experience integrating Spark with third-party systems and APIs to streamline data workflows.
    • Front-End Development: Basic to intermediate skills in front-end development to collaborate on building monitoring dashboards for Spark systems.
    • Software Development: Strong programming fundamentals, experience with version control (Git), and a solid understanding of CI/CD pipelines.


    Preferred Qualifications:


    • Experience with Kubernetes for managing Spark clusters in containerized environments.
    • Familiarity with cloud platforms such as AWS, GCP, or Azure.
    • Knowledge of SQL and database integration with Spark.
    • Experience with big data tools (e.g., Hadoop, Kafka) used alongside Spark.


    Why Join Us?


    • Be part of a forward-thinking, tech-driven team.
    • Work on cutting-edge distributed systems using federated models.
    • Collaborate with experts in cloud computing, big data, and data engineering.
    • Opportunities for professional growth and continuous learning.

    Check similar offers

    Senior Data Engineer z j. angielskim lub niemieckim (People and Project Analytics)

    New
    dmTECH Polska
    13.4K - 18.3K PLN
    Łódź
    , Fully remote
    Fully remote
    Python
    Java
    Snowflake

    Senior BI - Data Engineer

    New
    Antal Sp. z o.o.
    30K - 40K PLN
    Katowice
    , Fully remote
    Fully remote
    AI
    Databricks
    Azure

    Data Engineer (Azure Cloud Data Platform)

    New
    Baloise Solution Hub
    22K - 36K PLN
    Warszawa
    , Fully remote
    Fully remote
    Azure Data Factory
    MS Fabric
    Azure Synapse

    Senior Data Engineer (100% remote)

    New
    Crestt
    17K - 24K PLN
    Kraków
    , Fully remote
    Fully remote
    Power BI
    SQL
    Python

    Cloud Data Architect (AWS)

    New
    Future Processing
    24.8K - 36.8K PLN
    Wrocław
    , Fully remote
    Fully remote
    CI/CD
    Data Lake
    AWS