




**Data Engineer – Master \& Reference Data** **Location:** Mexico (Remote Mexico) **Employment Type:** C2H **Job Summary** We are seeking a highly motivated **Data Engineer** to join a Data Governance team, with a strong focus on **Master Data Management (MDM)** and **Reference Data Management**. In this role, you will help design, develop, and maintain our internal data platform to ensure consistency, integrity, and high\-quality master data across the organization. The ideal candidate is a proactive self\-starter, passionate about data, highly analytical, and committed to driving data excellence. **Key Responsibilities** * Lead the requirements gathering, design, development, and testing of master data solutions across multiple data domains. * Design, develop, and maintain master and reference data models within the Data Lakehouse. * Build and optimize batch and streaming data pipelines to ensure reliable data flows. * Collaborate with cross\-functional teams to align data strategies, processes, policies, and controls. * Integrate data from various sources into the Data Lakehouse and manage data synchronization into the data hub via API endpoints. * Implement and monitor data quality standards; investigate and resolve data exceptions with data stewards. * Conduct data profiling, source system analysis, and data mapping as needed. * Support workflow improvements and change management initiatives. * Share MDM best practices and provide training and guidance to data stewards. * Document data flows, lineage, catalogs, standards, and procedures to enhance data literacy across the organization. **Required Skills \& Experience** * **4\+ years of experience** as a Data Engineer, including **2\+ years working specifically with master and reference data**. * Strong expertise in **Master Data Management**, dimensional modeling, and building scalable MDM solutions. * Proficiency in **SQL, Python, Spark**, and modern data modeling practices. * Hands\-on experience with cloud data platforms such as **Databricks**, AWS **RDS**, and REST APIs. * Familiarity with engineering best practices, CI/CD workflows (e.g., GitHub), and code reviews. * Working knowledge of Agile/Scrum methodologies and tools (e.g., Jira, Confluence). * Understanding of **Generative AI** concepts and their application in data engineering. * Strong business acumen across multiple data domains—experience in **logistics** is a plus. * High attention to detail with strong analytical and problem\-solving skills. * Excellent verbal and written communication skills. * Ability to work independently, learn quickly, and adapt to changing priorities. * Experience with MDM tools such as Informatica or Stibo is **not required**, but hands\-on experience with master data and slowly changing dimensional models in **Databricks** is essential. * Ability to manage a custom\-built master data hub (AWS RDS \+ API Gateway). * Background in **Data Analysis, Statistics, or Business Analytics**. * Familiarity with the **AWS Cloud Data Ecosystem**. Job Type: Contract Contract length: 12 months Pay: $45,000\.00 \- $60,000\.00 per month Application Question(s): * Do you have at least 4 years of hands\-on experience working specifically with Master Data or Reference Data? * Please describe your experience working with Databricks, Python, SQL, and Spark. Include the types of data pipelines or models you have built. * Explain your experience designing or managing dimensional data models, including any work with slowly changing dimensions (SCD). Language: * English (Required) Work Location: Remote


