About the role:
We’re looking for a hands-on Data Engineer to join us on a 12-month max-term contract and play a key role in supporting and scaling our growing data infrastructure.
In this role, you’ll be responsible for building and maintaining scalable ETL/ELT pipelines using Databricks and modern cloud tools. You’ll also step in to temporarily support our business intelligence needs, developing and maintaining reports and dashboards in ThoughtSpot (or a similar BI platform).
You’ll collaborate closely with our Lead Data Engineer, who will provide architectural guidance and help drive the strategic direction of our data transformation initiatives.
This role is a great fit for a data engineer who enjoys working across the full data stack—from raw data ingestion and transformation all the way to the BI layer—with a strong focus on data quality, reliability, and usability.
We offer a hybrid work arrangement: 3 days in the office and 2 days remote each week, giving you the flexibility to do your best work.
Key Responsibilities:
Data Engineering
* Build, maintain, and optimize robust ETL/ELT pipelines using Databricks.
* Contribute to the design and implementation of data lake and data warehouse architecture.
* Translate business requirements into reliable and scalable data solutions.
* Collaborate with the Lead Data Engineer on data modeling, pipeline design, and cloud infrastructure best practices.
* Implement monitoring, alerting, and logging for data pipelines to ensure data integrity and reliability.
* Participate in sprint planning, technical documentation, code reviews, and team collaboration rituals.
BI & Reporting Support
* Maintain and support dashboards and reports in ThoughtSpot.
* Assist stakeholders with ad hoc data queries and visualization needs.
* Ensure availability and accuracy of key business metrics during the analyst’s leave period.
* Translate complex datasets into usable, decision-support insights.
Key Requirements:
Essential
* Strong experience building and managing ETL/ELT pipelines in Databricks or similar platforms.
* Proficiency in Python and SQL for data processing, transformation, and analysis.
* Deep knowledge of data modeling and warehousing concepts.
* Experience with BI tools, preferably ThoughtSpot (or Power BI, Tableau, Looker).
* Solid version control and collaboration practices (e.g., Git).
* Ability to collaborate closely with both technical and non-technical team members.
* Effective communication and problem-solving skills.
Desirable
* Exposure to DevOps practices such as CI/CD (e.g., Azure DevOps), infrastructure as code (e.g., Terraform).
* Experience working in a remote or distributed team environment.
* Experience working with cloud environments (AWS, Azure, or GCP).
* Familiarity with AWS services like S3, Lambda, EC2, SQS, and SNS.
Personal Attributes
* Curious, proactive, and eager to learn.
* Comfortable balancing engineering depth with occasional BI support.
* Strong ownership mindset and attention to detail.
* Organized and efficient; able to manage priorities across domains.
If you would like to be a part of the Wilbur team then please submit your application. We look forward to hearing from you!
#J-18808-Ljbffr