Singapore
Onsite
IT
Key Responsibilities
Design, build, and maintain scalable data pipelines for batch and real-time processing using Python, SQL, ETL/ELT frameworks, and big-data technologies.
Participate in end-to-end data project delivery using SDLC, Agile, or hybrid development methodologies.
Work closely with business and technology stakeholders to understand data requirements related to banking products, transactions, customer analytics, and regulatory reporting.
Develop efficient normalized and de-normalized data models for operational and analytical workloads.
Design and manage data warehouses, data marts, and integration layers aligned with enterprise data architecture.
Deploy physical data models and optimize performance for large-scale financial datasets.
Ensure adherence to data governance, quality, metadata, and privacy standards across all solutions.
Produce and maintain data documentation including dictionaries, lineage diagrams, and technical specifications.
Support data lineage, metadata management, and data quality initiatives to improve transparency and trust.
Provide data-driven assistance to business users and proactively communicate technical challenges.
Present insights, designs, and concepts effectively to both technical and non-technical stakeholders.
Skills/Experience:
Technical Skills
Proficient in Python; experienced with Spark for scalable ETL/ELT pipelines.
Strong SQL experience with large-scale datasets and warehouse solutions.
Knowledge of Hadoop ecosystem tools such as Hive, Spark, and HDFS.
Experience with AWS services including Glue, Redshift, RDS, S3, and basic IAM/VPC/security configurations.
Hands-on Linux skills, shell scripting, and AWS CLI usage.
Ability to work across SQL, NoSQL, and data lake environments.
Exposure to Terraform, Talend, or similar tools is a plus.
Familiarity with visualization tools such as QuickSight, Qlik, or Tableau.
Ability to write clean, production-grade code and maintain clear pipeline documentation.
Experience
Experience with large datasets on platforms such as Greenplum, Hadoop, Oracle, DB2, or MongoDB.
Familiarity with dashboarding tools (Tableau, Power BI, SAS VA).
Experience in scripting, application packaging, and deployment across DEV–PROD environments.
Understanding of change management, service request processes, and maintenance reporting.
Strong data modelling capabilities (logical and physical) for banking, risk, compliance, and analytics use cases.
Deep knowledge of relational/dimensional modelling, data warehousing concepts, and data integration techniques.
Strong SQL expertise supporting large and complex financial data environments.
Education \& Certifications
Bachelor’s degree in Software Engineering, Computer Science, or equivalent experience.
Professional cloud certifications (AWS/Azure/GCP) are preferred, including:
-
AWS Certified Data Analytics – Specialty
-
AWS Solutions Architect – Associate
-
Azure Data Engineer Associate
-
Google Professional Data Engineer
-
Databricks Data Engineer Associate/Professional
-
Cloudera Certified Data Engineer