ML Engineer
18 Month FTC
This role will require the software and dev ops capabilities of an ML Ops engineer to build ML and CI-CD pipelines, proficiency in Infrastructure as Code and Sagemaker pipelines combined with Data Engineering capabilities.
Role Summary:
Working across projects within LNER and the wider industry through our parent company DFTO the ML Ops Engineer will have responsibility for:
1. Designing and implementing automation pipelines to operationalise the ML platform and ML pipelines for CI/CD Pipelines
2. Responsibility for mapping out data feeds into / out from the ML platform in collaboration with Solutions Architects and the IT Team
3. Ensuring that the ML systems are developed and scaled reliably in line with the best practices for ML Ops and software engineering, ethics, and system security
4. Creation of Data Engineering pipelines utilising Infrastructure as code (IAC) to ensure projects ingest data in an automated way that is quality checked and fit for purpose for machine learning modelling.
Knowledge and Skills of Candidate:
Essentials
5. Proficient in Unix environment and scripting in Bash and Python.
6. Strong experience with AWS infrastructure with proficiency in the services such as: S3, EC2, Lambda functions, Cloud Formation, Athena, Dynamo DB, Code Commit, SageMaker, etc.
7. Strong experience with containerization using Docker and containers management.
8. Strong software engineering skillset including code review: a good understanding of coding best practices and experience with code and data versioning (using Git/CodeCommit), code quality and optimisation, error handling, logging, monitoring, validation and alerting.
9. Fluent in writing well tested, readable code that is capable of processing large volumes of data and large amount of data processing and ML jobs, including the use of IAC to build data pipelines
10. Expert knowledge of Python.
11. An excellent knowledge of basic machine learning libraries, such as NumPy, SciPy, Pandas, Dask, PyTorch, Tensorflow, etc.
12. A proven track record of linking data from multiple systems for scalable productionised solutions with security and monitoring best practices.
13. Experienced with Cloud Security best practices.
14. Hands-on experience with DevOps lifecycle, tools and frameworks.
Desirables
15. Knowledge of ML approaches such supervised/unsupervised machine learning, reinforcement learning, Bayesian inference.
16. AWS Certification is strong benefit.
17. Experience with Google Cloud’s Big Data tools.
18. Proficient with Kafka is plus but not essential.