About the Role
We are searching for a
skilled Data Engineer
to join our team in Mexico City. The ideal candidate will have extensive hands-on experience with
Azure Databricks and Spark
, strong coding proficiency in
Python (PySpark)
and
SQL
, and a deep understanding of
data lake architecture
,
ETL processes
, and
data quality management
.
Experience with
Azure data analytics services
and excellent
technical writing skills
are also required.
Key Responsibilities
Develop and implement robust data processing pipelines using
Azure Databricks
and
Apache Spark
.
Write efficient and optimized
Python (PySpark)
and
SQL
code for
terabyte-scale
data processing.
Design, build, and maintain
data lake architectures
and
ETL workflows
.
Implement strong
data quality and validation frameworks
to ensure accuracy and reliability.
Leverage
Azure data analytics services
for data storage, transformation, and reporting.
Prepare detailed
technical documentation
, including requirements and testing specifications.
Collaborate with cross-functional teams to align on data requirements and deliver scalable solutions.
Manage source code and version control using
Git
.
Qualifications \& Skills
5+ years
of experience in development and production implementation with
Azure Databricks
and
Apache Spark
.
Advanced proficiency in
Python (PySpark)
and
SQL
.
Proven ability to
optimize large-scale (TB-level) data pipelines
.
Strong knowledge of
data lake architecture
,
ETL processes
, and
data quality frameworks
.
Hands-on experience with
Azure Data Analytics Services
(e.g., Data Factory, Synapse, ADLS).
Excellent
technical documentation and communication skills
.
Proficient in
Git
for version control and collaboration.