Job Title: Senior Data Engineer
Experience:
6 - 8 years
Location:
Hybrid
Employment Type:
Full-time
Mandatory skills: AWS, Python, SQL, Databricks
About The Role
We are looking for a highly skilled
Senior Data Engineer
who is passionate about building robust, scalable, and high-performance data systems. The ideal candidate will have deep expertise in
SQL, Python, AWS, and Databricks
, with a proven track record of designing and implementing modern data pipelines and analytical frameworks.
Key Responsibilities
Design, develop, and maintain scalable data pipelines and ETL processes for data ingestion, transformation, and storage.
Work with cross-functional teams to define and deliver data solutions supporting business and analytics needs.
Optimize and fine-tune SQL queries, data models, and pipeline performance.
Build and manage data workflows in Databricks and integrate with AWS data services (S3, Redshift, Glue, Lambda, etc.).
Ensure data accuracy, consistency, and reliability through data quality checks and monitoring frameworks.
Collaborate with Data Scientists, Analysts, and Product teams to enable self-service analytics and advanced data-driven insights.
Follow best practices for data governance, security, and compliance.
Continuously evaluate emerging data technologies and propose innovative solutions for process improvement.
Required Skills \& Qualifications
Bachelor’s or master’s degree in computer science, Information Technology, or a related field.
5+ years of hands-on experience in Data Engineering or related roles.
Strong proficiency in SQL for complex query development and data manipulation.
Expertise in Python for building data processing and automation scripts.
Experience with AWS ecosystem — especially S3, Glue, Redshift, Lambda, and EMR.
Hands-on experience with Databricks for data processing, transformation, and analytics.
Experience working with structured and unstructured datasets in large-scale environments.
Solid understanding of ETL frameworks, data modeling, and data warehousing concepts.
Excellent problem-solving, debugging, and communication skills.
Good to Have
Experience with Airflow, Snowflake, or Kafka.
Knowledge of CI/CD pipelines and Infrastructure as Code (IaC) tools such as Terraform or CloudFormation.
Exposure to data governance, metadata management, and data cataloguing tools.
Skills: python,databricks,sql,aws