Salary: £100,000 - 100,000 per year Requirements: Bachelors or Masters degree in Computer Science, Electrical Engineering, or related field Strong knowledge of distributed systems, operating systems, and machine learning systems architecture Experience with inference serving and AI infrastructure Hands-on experience with LLM serving frameworks (e.g., vLLM, Ray Serve, TensorRT-LLM, TGI) Knowledge in distributed KV cache optimization Proficiency in C/C++, with additional experience in Python for research prototyping Solid grounding in systems research methodology, distributed algorithms, and profiling tools Team-oriented mindset with effective technical communication skills PhD in systems, distributed computing, or large-scale AI infrastructure (desired) Publications in top-tier systems or ML conferences (desired) Responsibilities: Collaborate with senior architects on real-world projects Prototype and optimize next-generation AI infrastructure Conduct research to improve performance of distributed AI systems Design and implement solutions for distributed systems and machine learning workflows Analyze and refine system architectures for efficiency and scalability Communicate findings and strategies effectively within the team Technologies: AI LLM Machine Learning Python Cloud More: We are a leading research-driven organization focused on advancing computer systems and AI infrastructure. Our team is composed of innovative thinkers dedicated to solving complex challenges in distributed AI and performance optimization. We offer competitive benefits, opportunities for professional growth, and a collaborative work environment. This position is based in a dynamic location where cutting-edge technology is at the forefront of our projects. last updated 6 week of 2026