Description and Requirements
This role is open for the Edinburgh, Scotland location only. Candidates must be based there, as the position requires working from the office at least three days per week (3:2 hybrid policy).
The Lenovo AI Technology Center (LATC)—Lenovo’s global AI Center of Excellence—is driving our transformation into an AI-first organization. We are assembling a world-class team of researchers, engineers, and innovators to position Lenovo and its customers at the forefront of the generational shift toward AI. Lenovo is one of the world’s leading computing companies, delivering products across the entire technology spectrum, spanning wearables, smartphones (Motorola), laptops (ThinkPad, Yoga), PCs, workstations, servers, and services/solutions. This unmatched breadth gives us a unique canvas for AI innovation, including the ability to rapidly deploy cutting-edge foundation models and to enable flexible, hybrid-cloud, and agentic computing across our full product portfolio. To this end, we are building the next wave of AI core technologies and platforms that leverage and evolve with the fast-moving AI ecosystem, including novel model and agentic orchestration & collaboration across mobile, edge, and cloud resources. This space is evolving fast and so are we. If you’re ready to shape AI at a truly global scale, with products that touch every corner of life and work, there’s no better time to join us.
We are seeking a highly motivated and skilled AI Cloud Engineer to join our rapidly growing AI team. You will play a critical role in the training of large language models (LLMs), large vision models (LVMs), and large multimodal models (LMMs), including fine-tuning and reinforcement learning. This is a challenging yet rewarding opportunity to contribute to cutting-edge research and development in generative AI. You’ll be working with a collaborative team to push the boundaries of what’s possible with AI models and deploy them into innovative products. If you are passionate about making Smarter Technology For All, come help us realize our Hybrid AI vision!
Responsibilities:
1. Design, implement, and evaluate training pipelines for large generative AI models, encompassing multiple stages of post-training.
2. Data augmentation: Design, implement, and evaluate data augmentation pipelines to increase the diversity and robustness of training datasets, improving model performance, particularly in low-data regimes.
3. Adversarial training: Develop and implement adversarial training techniques to improve model robustness against adversarial attacks and enhance generalization performance by exposing the model to perturbed input examples during training.
4. Supervised Fine-tuning (SFT): Developing and executing SFT strategies for specific tasks.
5. Reinforcement Learning from Human Feedback (RLHF): Running and refining RLHF pipelines to align models with human preferences.
6. Pruning: Design and implement model pruning strategies to reduce model size and computational complexity by removing non-essential parameters, optimizing for both performance and efficiency without significant accuracy loss.
7. Distillation: Develop and perform model distillation techniques to compress large language models into smaller, more efficient models while preserving key performance characteristics.
8. Quantization: Implement and evaluate model quantization techniques to reduce model size and accelerate inference speed, balancing precision loss with performance gains for deployment across diverse hardware platforms.
9. Low-Rank Adaptation (LoRA): Utilizing techniques for efficient fine-tuning of large language models, balancing performance and resource constraints, and tailoring model performance for downstream tasks well.
10. Experiment with various training techniques, hyperparameters, and model architectures to optimize performance and efficiency.
11. Develop and maintain data pipelines for processing and preparing training data.
12. Monitor and analyze model training progress, identify bottlenecks, and propose solutions.
13. Stay up-to-date with the latest advancements in large language models, training techniques, and related technologies.
14. Collaborate with other engineers and researchers to design, implement, and deploy AI-powered products.
15. Contribute to the development of internal tools and infrastructure for model training and evaluation.
Qualifications:
16. Bachelor's or Master's degree in Computer Science, Machine Learning, or a related field and 5+ years of relevant work experience or 7+ years of relevant work experience.
17. Strong programming skills in Python and experience with deep learning frameworks like PyTorch.
18. Solid understanding of machine learning principles, including supervised learning, unsupervised learning, and reinforcement learning.
19. Proven experience in designing and conducting experiments, analyzing data, and drawing meaningful conclusions.
20. Familiarity with large language models, transformer architectures, and related concepts.
21. Experience with data processing tools and techniques Pandas, NumPy).
22. Experience working with Linux systems and/or HPC cluster job scheduling Slurm, PBS).
23. Excellent communication, collaboration, and problem-solving skills.
Bonus in Computer Science, Machine Learning, or a related field.
24. Experience with distributed training frameworks DeepSpeed, Megatron-LM).
What we offer:
25. Opportunities for career advancement and personal development
26. Access to a diverse range of training programs
27. Performance-based rewards that celebrate your achievements
28. Flexibility with a hybrid work model (3:2) that blends home and office life
29. Electric car salary sacrifice scheme
30. Life insurance
#LATC #LATC-IC