We are looking for an Inference Platform Engineer (LLM & Kubernetes) to join our team.
Our client is a leading European AI company developing large language models and generative platforms for enterprise and government clients. Their products combine high-performance technologies, transparency, accessibility, and data security, fully aligned with European regulatory and ethical standards.
As an Inference Platform Engineer (LLM & Kubernetes), you will take ownership of inference API integration, operations, and platform reliability across production AI systems. This role is designed to be covered by 1–2 FTE split across several senior specialists, ensuring continuity of inference services and full coverage during planned and unplanned absences as we take over end-to-end LLM inference responsibility.
We offer:
*not applicable for freelancers