We are looking for a highly skilled and specialized
Data Engineer
Job Title:
Data Engineer
Job Location:
Toronto
Job Type:
Full - Time
This is an onsite position requiring employees to work from the office five times per week.
This role requires a unique blend of expertise spanning established open-source tools
(Talend, Pentaho), modern cloud-native platforms (Azure Data Factory, Airbyte
), and deep technical data management skills including advanced data modeling, SQL mastery, and rigorous STTM documentation.
The ideal candidate will bridge legacy systems and modern cloud architectures within a demanding, data-intensive environment.
Key Responsibilities
Design, build, automate, and support robust ETL pipelines utilizing a hybrid stack:
Orchestrate and execute data movement within the Azure ecosystem.
Manage and modernize existing on-premise and open-source data flows.
Implement and customize connectors for rapid SaaS/API integration.
Design and implement highly optimized data warehouse schemas (star schema, snowflake schema, data vaults) to support analytical reporting and business intelligence needs.
Drive the requirements gathering process, creating detailed, rigorous STTM documents that clearly define data transformations, data quality rules, and metadata definition
Write complex SQL, T-SQL, or PL/SQL queries and stored procedures, and perform expert-level performance tuning on large-scale relational databases (e.g., SQL Server, Azure SQL DB, Snowflake).
Implement rigorous data validation, cleansing, and error handling mechanisms across all platforms to ensure data integrity and compliance with industry standards.
Leverage Azure services (Data Lake Storage Gen2, Synapse Analytics, etc.) in conjunction with ETL tools to build scalable cloud-native data solutions.
Use scripting languages (Python, PowerShell) to automate job orchestration, monitoring, and testing procedures not natively covered by the ETL tools.
-
Required Skills and Qualifications
10+ years of dedicated experience in ETL development and data warehousing, specifically working with multiple technologies mentioned above in an enterprise setting.
Familiarity with BI tools and specific technologies like
SSAS (SQL Server Analysis Services), MDX, and OLAP cubes
Experience with big data frameworks like
Hadoop, Spark, and related components such as HDFS, Hive, and Sqoop
Experience in scripting languages like
Python, Java, or Scala
for automation and complex data manipulation
Understanding how to build and manage data pipelines that feed machine learning models and AI applications.
Hands-on experience building pipelines, managing linked services, datasets, and triggers.
Experience in using
Python/Pandas
for custom, production-ready ETL pipelines
Experience managing and developing jobs within the open-source frameworks like
Talend \& Pentaho (PDI).
Familiarity with deployment and connector utilization/customization
using Airbyte.
Proven experience designing dimensional and normalized data models.
Mastery of complex SQL query writing and performance optimization techniques.
Demonstrated ability to produce detailed and accurate Source-to-Target Mappings and functional specifications.
Experience with Python or PowerShell scripting.
Bachelor's degree in Computer Science, Information Technology, or a related quantitative field.
Strong analytical ability, meticulous attention to detail, excellent communication skills, and the capacity to translate complex technical concepts into clear documentation for stakeholders.
Preferred Qualifications
Microsoft Certified: Fabric Data Engineer Associate - DP700
Microsoft SQL Certification
Microsoft Certified Professional(MCP)
Microsoft Certified IT Professional (MCITP)
Experience with workflow orchestration tools like Apache Airflow.
Knowledge of Big Data platforms or NoSQL databases.
If interested, please apply through LinkedIn or share your resumes at hr@cloudvital.ca