About This Role
Location: Hybrid- 3 days a week from our Winnipeg or Toronto office 2 days working from home Reporting to: Senior Team Lead - Data Engineering
-
Implement and monitor data quality rules and checks using SQL and Python within our GCP environment, primarily leveraging BigQuery and Cloud Storage.
-
Assist in managing data access controls and permissions under the guidance of established governance policies.
-
Support the implementation and monitoring of data governance policies, including data lineage tracking, data security classifications, and metadata management.
-
Contribute to the creation and maintenance of clear and concise documentation for data governance standards, processes, and platform usage.
-
Develop and maintain Python scripts to automate data governance tasks and streamline platform operations.
-
Collaborate effectively with platform engineers and data teams to identify, troubleshoot, and resolve data governance-related issues.
-
Utilize Git for version control of code and configurations, ensuring organized and collaborative development.
-
Actively engage in learning about data governance best practices, GCP data services, and the overall data platform architecture through mentorship and independent study.
-
Foundational understanding of core data concepts, including data warehousing principles (like schema design), ETL/ELT processes, data modeling fundamentals, and the importance of data quality.
-
Basic SQL skills with the ability to write, understand, and debug queries for data manipulation, transformation, and basic analysis.
-
Basic proficiency in Python for scripting and automation of tasks.
-
Familiarity with cloud fundamentals, specifically Google Cloud Platform (GCP) and its core data services such as BigQuery and Cloud Storage. A basic understanding of IAM principles within GCP is also important.
-
Awareness of fundamental data governance concepts, including data quality, data lineage, data security, and metadata management.
-
Demonstrated problem-solving and analytical skills with a keen attention to detail.
-
Effective written and verbal communication skills, with the ability to document your work and collaborate with team members.
-
Basic understanding of Git for version control and collaborative coding.
-
A strong eagerness to learn and grow your expertise in data governance and the data platform domain.
Nice-to-Have Qualifications
- Exposure to other GCP data services like Cloud Dataflow, Cloud Composer (Apache Airflow on GCP), Data Catalog, and Cloud Functions/Cloud Run.
- Familiarity with concepts or tools related to data quality management or data lineage tracking.
- Basic understanding of data governance frameworks (e.g., GDPR, CCPA) or data masking/anonymization techniques.
- Basic understanding of Infrastructure as Code (IaC) tools such as Terraform or Deployment Manager.
- Basic familiarity with data visualization tools (e.g., Looker, Tableau, Google Data Studio).
- Experience interacting with RESTful APIs to retrieve or send data.
- Familiarity with other programming languages such as Java or Scala.
- A foundational understanding of basic statistical concepts.
At Skip, This Is On The Menu