Software Developer
Permanent | Hybrid (willing to travel to Newcastle) | Python | Apache | Data | SC cleared
Hybrid work arrangement. Office attendance is up to 60%. Location is flexible: London, Leeds, Newcastle, Sheffield, Blackpool, Manchester and Birmingham. You must be willing to travel to Newcastle when required.
The Role:
We are looking for Software Developers with strong Python and data processing with Apache Spark responsible to lead the design, development, and operation of data-driven applications and pipelines in a collaborative Dev and DevOps environment.
The role focuses on writing and improving application code while working closely with DevOps engineers to support automated deployments, infrastructure management, system monitoring, reliability, and scaling. The post holder collaborates with Product Owners, Business Analysts, and users to translate business needs into robust technical solutions, operates and improves production services, analyses data and system issues to identify root causes, and champions engineering best practices. They also provide technical leadership through coaching and mentoring junior colleagues and contributing to continuous improvement across development, delivery, and operational processes.
Responsibilities:
Engineers will contribute to research, development and delivery across:
Design, build, and operation of data ingest and publishing pipelines
Workflow orchestration and task scheduling using managed services
Collaboration with Product Owners, Business Analysts, and users to shape technical solutions
Production support, monitoring, and continuous improvement of system resilience, stability, and performance
Data analysis and investigation to identify root causes of defects and operational issues
DevOps collaboration, including supporting automated deployments, infrastructure management, monitoring, and scaling
Coaching and mentoring junior engineers and promoting engineering best practices
Skills & Experience:
Understanding of data processing using Apache Spark
Use of Python, SQL, and familiarity with PySpark
Experience using Apache Airflow for task orchestration
Understanding of EMR and reviewing output logs
Use of Jupyter notebooks and/or Amazon Athena to query and validate data
Data analysis to identify root cause of issues
Understanding of dimensional data models and slowly changing dimensions/historic data capture
Use of AWS console and services such as, but not limited to; CloudWatch, IAM, S3, Glue, ECR, EC2, EMR, Dynamo DB, LakeFormation
Familiarity with Amazon Textract and Comprehend
Understanding of both server-side and client-side encryption
Use of GitLab for source code management pipelines for CI/CD
Use of GitLab Tags for component versioning in shared repositories
Understanding of Docker and containerization of solutions
IaC using Terraform
Experience of understanding how customer expectations transition to applied functionality
Familiarity with, and implementation of Engineering best practices
Use of gitlab for release tagging and deployments
Familiarity with basic data structures for constructing a solution
Active BPSS, SC clearance or eligible for clearance
Desirable skills :
Experience supporting AI or data driven platforms
Knowledge of cyber security or fraud prevention domains
Experience working within government or critical national infrastructure environments
Find out more: peregrine.global or check out our LinkedIn page: peregrine-resourcing
TPBN1_UKTJ