Job Description
PySpark Developer (SAS & AWS)12+ Month Contract | UK Fully Remote | Inside IR35
We're supporting a large-scale data and AI transformation programme. We are looking for an experienced Data & AI - LLM Model Developer with advanced PySpark and AWS expertise to help modernise complex Legacy data platforms.
This is a hands-on contract role where you'll play a key part in SAS-to-PySpark migration, building scalable, cloud-native data pipelines and delivering production-ready solutions in a regulated environment.
What you'll be doing
* Designing, developing, and optimising PySpark-based data pipelines on AWS
* Converting Legacy SAS workloads to PySpark, using automated migration tools and manual optimisation
* Refactoring and stabilising existing data workflows into modern cloud architectures
* Optimising Spark workloads for performance, scalability, and cost efficiency
* Working closely with engineers and stakeholders to deliver reliable, high-quality data solutions
Essential skills
* PySpark: strong hands-on experience building production-grade Spark solutions
* AWS: EMR, Glue, S3, Athena (and related services)
* Experience using automated migration tools for large-scale code or data modernisation
* Strong SQL and data engineering fundamentals
* Experience working with distributed data processing and cloud platforms
Nice to have
* Exposure to SAS or Legacy analytics platforms
* Experience in banking or financial services
* CI/CD, Git-based workflows, or DevOps tooling
Why this role?
* Fully remote (UK-based) contract
* Long-term transformation programme with real technical depth
* Modern cloud and data stack
* Outside-the-box problem solving, not just maintenance work