We are seeking a talented Senior Edge AI Engineer/Architect with specialized expertise in GPU/TPU acceleration to join our team. The ideal candidate will have extensive hands-on experience in local Large Language Models (LLM) inference with embedded GPU/TPU architectures. As Principal Engineer specializing in Edge AI, you will play a crucial role in shaping the future Edge AI solution, leveraging the power of GPU/TPU acceleration and enterprise grade, large scale edge compute.
The successful candidate will combine technical excellence with effective leadership, creating a positive impact on both projects and team dynamics.
Key Responsibilities:
High-Level Design and Architecture
- Influence the Edge AI strategy by providing expert advice on design and architecture
- Make critical decisions regarding technical directions, scalability, and system performance
- Develop and optimize AI inference models for deployment on edge devices with embedded GPU/TPU accelerators, focusing on local Low Latency Model (LLM) inference
- Implement and fine-tune low-latency model inference pipelines to meet real-time performance requirements
- Collaborate with cross-functional teams to integrate AI inference solutions into edge computing platforms and applications
- Collaborate with the GPU Hardware Design Team to design and optimize GPUs that power next-generation devices
- Conduct performance profiling and optimization to maximize the efficiency of GPU/TPU acceleration for local LLM inference
- Work on micro-architecture development, ensuring efficient execution of graphics, compute, and AI workloads within energy and area constraints
- Stay current with advancements in GPU/TPU technologies and edge AI frameworks, incorporating them into solution designs as appropriate
- Provide technical expertise and support to project teams, ensuring successful implementation and deployment of edge AI solutions
Team Leadership:
- Lead and inspire a team of engineers, providing guidance, setting goals, and ensuring collaboration
- Oversee project planning, execution, and delivery, ensuring alignment with business objectives
- Manage all phases of technical projects, from conception to completion
- Develop project specifications, track progress, and control costs
- Foster a positive work environment, encouraging professional growth and knowledge sharing
Qualifications:
- Bachelor’s degree in computer science, Engineering, or a related field; master’s degree preferred
- 5+ years of hands-on experience in AI model development and deployment, with a focus on edge computing and local LLM inference
- Strong programming skills in languages such as Python and C++
- Proficiency in LLM frameworks (e.g., vLLM, Text generation inference, OpenLLM, Ray Serve, and HuggingFace Transformers) and deep learning libraries
- Extensive experience with GPU/TPU acceleration for AI inference, including optimization techniques (tensor, pipeline, data, sharded data parallelism) and performance tuning,
- Hands on experience with one or more GPU frameworks: CUDA, Vulkan, OpenCL
- Deep knowledge of GPU memory layout, familiarity with NVIDIA Jatison, ARM Mali or relevant SoC configurations
- Knowledge of parallel computation, memory scheduling, and structural optimization
- Excellent problem-solving and analytical skills, with a passion for innovation and continuous learning
Additional Skills (Preferred):
- Experience with edge device hardware and software integration
- Familiarity with edge computing architectures and IoT platforms
- Experience with edge AI applications in domains such as robotics, autonomous vehicles, or industrial automation
Benefits:
- Private healthcare
- Multisport card
- Flat structure
- In-house trainings
- Small teams
- Perks in the office (Bike parking, Shower, Playroom)
- No dress code
Who we are as a team
- We have a strong sense of responsibility to our clients and the quality of our work. We are accountable for the projects, we tackle clients’ challenges and we go the extra mile to deliver
- Clients praise us for flexibility. If something is humanly possible, we do it
- We are fair, honest and transparent with each other and with our clients
- We strive for top quality and don’t settle for “standards”
- We trust each other, work well together and deliver value to clients as a team
Who we are as as company
R Systems specializes in digital transformation services based on the latest technologies in the area of cloud , data analytics, RPA, IA, ML/AI. For almost 30 years in the telecom and technology industry, we have successfully delivered over 1000 complex, high-stake projects to more than 100 operators worldwide, supporting them throughout the 2G to 5G evolution. As we progress into the digital transformation era, we aim to become the tech partner of choice for global organizations looking to innovate and expand their business through the latest technology.
-
2,700 employees worldwide
-
25 offices in 12 countries in Europe, US, APAC
-
16 delivery hubs, including centers of excellence in Romania, Poland, Moldova
-
Partners of Microsoft– cloud computing and cloud application strategic partner; UiPath – RPA/IA Silver partner; Google Cloud; Oracle – Gold Partner; Automation Anywhere and Salesforce – US Gold Partner