Overview
I am supporting a university with a major data platform transformation project as they implement AWS across their environment. We are looking for a Data Engineer with strong hands‑on experience in designing and delivering enterprise‑scale data pipelines using AWS Glue and PySpark. The role will involve building and optimising ETL processes, working with raw and curated datasets, and ensuring data is processed efficiently and to a high standard.
Responsibilities
You will be responsible for developing scalable, production‑grade data workflows, integrating data from multiple systems, and applying best practices around data modelling, data quality, and automation. Experience working within a modern cloud data stack is essential, along with an understanding of how to structure data for analytics, reporting and downstream consumption.
Desired Skills and Experience
The ideal candidate will have a solid background in Spark‑based engineering, particularly PySpark, and be confident working with Glue jobs, Glue Catalog, S3, and other AWS native services used within a data platform build. Proven ability to build and optimise ETL processes, integrate multiple data sources, and uphold data quality and modelling best practices in a cloud environment.
Location and Terms
Location: Remote (client based in North East England)
Rate: £500‑£600 per day
IR35: Inside IR35, must use an approved umbrella on our list
Duration: approx 3 months
Start date: ASAP
#J-18808-Ljbffr