As the world's leading and most diverse derivatives marketplace, CME Group (www.cmegroup...
Accountabilities
* Automate data tasks on Google Cloud Platform (GCP).
* Work with data domain owners, data scientists, and other stakeholders to ensure that data is consumed effectively on GCP.
* Design, build, secure, and maintain data infrastructure, including data pipelines, databases, data warehouses, and data processing platforms on GCP.
* Measure and monitor the quality of data on GCP data platforms.
* Implement robust monitoring and alerting systems to proactively identify and resolve issues in data systems.
* Respond to incidents promptly to minimize downtime and data loss.
* Develop automation scripts and tools to streamline data operations and make them scalable to accommodate growing data volumes and user traffic.
* Optimize data systems to ensure efficient data processing, reduce latency, and improve overall system performance.
* Collaborate with data and infrastructure teams to forecast data growth and plan for future capacity requirements.
* Ensure data security and compliance with data protection regulations.
* Implement best practices for data access controls and encryption.
* Collaborate with data engineers, data scientists, and software engineers to understand data requirements, troubleshoot issues, and support data-driven initiatives.
* Continuously assess and improve data infrastructure and data processes to enhance reliability, efficiency, and performance.
* Maintain clear and up-to-date documentation related to data systems, configurations, and standard operating procedures.
Qualifications
* Education: Bachelor's or Master's degree in Computer Science, Software Engineering, Data Science, or a related field, or equivalent practical experience.
* Professional Experience: Experience as a Site Reliability Engineer or a similar role with a strong focus on data infrastructure management.
* Methodologies: Understanding of Site Reliability Engineering (SRE) practices.
* Data Technologies: Proficiency in data technologies such as relational databases, data warehousing, big data platforms (e.g., Hadoop), data streaming (e.g., Kafka), and cloud services (e.g., AWS, GCP, Azure).
* Programming Skills: Programming skills in languages like Python (NumPy, pandas, PySpark), Java (Core Spark with Java, functional interface, collections), or Scala with experience in automation and scripting.
* Infrastructure Tools: Experience with containerization and orchestration tools like Docker and Kubernetes is a plus.
* Compliance: Experience with data governance, data security, and compliance best practices on GCP.
* Software Development: Understanding of software development methodologies and best practices, including version control (e.g., Git) and CI/CD pipelines.
* Cloud Computing: Any experience in cloud computing and data-intensive applications and services, ideally Google Cloud Platform (GCP) would be highly beneficial.
* Quality Assurance: Experience with data quality assurance and testing on GCP.
* GCP Data Services: Proficiency with GCP data services (BigQuery, Dataflow, Data Fusion, Dataproc, Cloud Composer, Pub/Sub, Google Cloud Storage).
* Monitoring Tools: Understanding of logging and monitoring using tools such as Cloud Logging, ELK Stack, AppDynamics, New Relic, and Splunk.
* Advanced Tech: Knowledge of AI and ML tools is a plus.
* Certifications: Google Associate Cloud Engineer or Data Engineer certification is a plus.
Company Benefits
* Bonus Programme
* Equity Programme
* Employee Stock Purchase Plan (ESPP)
* Private Medical and Dental coverage
* Income Protection
* Life Assurance
* Cycle To Work
* Family Leave
* Education Assistance – MBA/Advanced Degree/Bachelor Degree
* Ongoing Employee Development Training/Certification
#J-18808-Ljbffr