Data Engineer – Master \& Reference Data
Location: Mexico (Remote Mexico)
Employment Type: C2H
Job Summary
We are seeking a highly motivated Data Engineer to join a Data Governance team, with a strong focus on Master Data Management (MDM) and Reference Data Management. In this role, you will help design, develop, and maintain our internal data platform to ensure consistency, integrity, and high-quality master data across the organization.
The ideal candidate is a proactive self-starter, passionate about data, highly analytical, and committed to driving data excellence.
Key Responsibilities
Lead the requirements gathering, design, development, and testing of master data solutions across multiple data domains.
Design, develop, and maintain master and reference data models within the Data Lakehouse.
Build and optimize batch and streaming data pipelines to ensure reliable data flows.
Collaborate with cross-functional teams to align data strategies, processes, policies, and controls.
Integrate data from various sources into the Data Lakehouse and manage data synchronization into the data hub via API endpoints.
Implement and monitor data quality standards; investigate and resolve data exceptions with data stewards.
Conduct data profiling, source system analysis, and data mapping as needed.
Support workflow improvements and change management initiatives.
Share MDM best practices and provide training and guidance to data stewards.
Document data flows, lineage, catalogs, standards, and procedures to enhance data literacy across the organization.
Required Skills \& Experience
4+ years of experience as a Data Engineer, including 2+ years working specifically with master and reference data.
Strong expertise in Master Data Management, dimensional modeling, and building scalable MDM solutions.
Proficiency in SQL, Python, Spark, and modern data modeling practices.
Hands-on experience with cloud data platforms such as Databricks, AWS RDS, and REST APIs.
Familiarity with engineering best practices, CI/CD workflows (e.g., GitHub), and code reviews.
Working knowledge of Agile/Scrum methodologies and tools (e.g., Jira, Confluence).
Understanding of Generative AI concepts and their application in data engineering.
Strong business acumen across multiple data domains—experience in logistics is a plus.
High attention to detail with strong analytical and problem-solving skills.
Excellent verbal and written communication skills.
Ability to work independently, learn quickly, and adapt to changing priorities.
Experience with MDM tools such as Informatica or Stibo is not required, but hands-on experience with master data and slowly changing dimensional models in Databricks is essential.
Ability to manage a custom-built master data hub (AWS RDS + API Gateway).
Background in Data Analysis, Statistics, or Business Analytics.
Familiarity with the AWS Cloud Data Ecosystem.
Job Type: Contract
Contract length: 12 months
Pay: $45,000.00 - $60,000.00 per month
Application Question(s):
Do you have at least 4 years of hands-on experience working specifically with Master Data or Reference Data?
Please describe your experience working with Databricks, Python, SQL, and Spark. Include the types of data pipelines or models you have built.
Explain your experience designing or managing dimensional data models, including any work with slowly changing dimensions (SCD).
Language:
English (Required)
Work Location: Remote