Software Engineer - Inference squad (AI Tribe)
Software Engineer - Inference squad (AI Tribe)
Software Engineer - Inference squad (AI Tribe)
2 days ago Be among the first 25 applicants
Get AI-powered advice on this job and more exclusive features.
OUR STORY: Join Scaleway and shape the sovereign cloud of tomorrow !
OUR STORY: Join Scaleway and shape the sovereign cloud of tomorrow !Since 1999, we have been designing secure, sustainable infrastructures aimed at supporting the most ambitious companies.Historically known for our dedicated servers (Dedibox), we made a strategic shift to cloud computing in 2015. Staying true to our principles of simplicity, flexibility, and technical excellence, we have become one of the leading players in Europe in the sector.With the rise of artificial intelligence, we have strengthened our commitment, supported by the Iliad Group, which is investing €3 billion to develop a serious, sovereign AI alternative to American and Asian giants.Every day, thanks to our rich catalog of products and services (bare metal, containerization, serverless, AI, etc.), Scaleway proudly serves 38,000 private and public sector clients, from Photoroom to Mistral AI, Golem AI, and ADEME.Our offices are located in Paris, Lille, Toulouse, Bordeaux, and Lyon. WHY WE NEED YOU?Our growth is driving us to strengthen our AI Infrastructure team to support our next-generation inference platform.Your mission will be to build and operate production-grade DevOps infrastructure for AI workloads, in order to enable the deployment and scaling of LLMs and generative AI applications on a sovereign European cloud.YOUR FUTURE TEAMWe work in a collaborative and international environment where the diversity of Scalers, combined with a spirit of sharing, helps bring new projects to life every day, advancing our ambitions together.You will be part of a team of 4 DevOps engineers, working closely on backend and infrastructure topics related to AI inference. The team is part of a broader AI tribe focused on two strategic products:
- Managed Inference, a platform to deploy, scale, and monitor AI models in production
- Generative APIs, a unified API layer to access cutting-edge generative models (LLMs, Diffusion, etc.)
- Build and operate infrastructure for serving LLMs and other generative models at scale
- Design, deploy, and maintain Kubernetes-based services optimized for AI inference
- Develop in Golang to build robust and efficient backend services
- Optimize serving stacks using tools like vLLM, Triton, and CUDA
- Integrate open-source AI tooling such as Hugging Face, KServe, or custom components
- Participate in architecture discussions and make high-impact technical decisions
- Keep up with the fast-paced evolution of AI serving ecosystems
- Troubleshoot and resolve complex infrastructure and deployment issues
- Collaborate with internal teams to integrate new AI features
- Contribute to code reviews and knowledge sharing within the team
Seniority level
Seniority level
Not Applicable
Employment type
Employment type
Full-time
Job function
Job function
Engineering and Information TechnologyIndustries
Software Development
Referrals increase your chances of interviewing at Scaleway by 2x
Sign in to set job alerts for “Software Engineer” roles.
Puteaux, Île-de-France, France 1 month ago
Software Engineer Intern - Backend or Fullstack
Software Engineer Intern - Front-End or Fullstack
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
#J-18808-Ljbffr