👨🏻‍💻 postech.work

SENIOR DATA ENGINEER (ROMA)

Dacomat • 🌐 In Person

In Person Posted 2 days, 8 hours ago

Job Description

Azienda

Dacomat srl

Posizione

Senior data engineer (roma)

Data udate

10/10/2025

Luogo lavoro

Roma

Impegno

Full-time

Compenso lordo

Da concordare

Posti disponibili

1

Contratto lavoro

Da determinare

Descrizione offerta di lavoro

Senior Data Engineer – Job Description

Location: Rome / Hybrid

Industry: Telecommunications / Technology

Reporting Line: Big Data General Manager

Overview

Our client, a leading multinational group in the telecommunications and technology sector, is seeking a highly skilled Senior Data Engineer to join its Big Data division. This strategic role is embedded within a high-impact team responsible for designing and operating scalable data platforms that support advanced analytics across global operations.

The ideal candidate will have extensive hands-on experience in designing, implementing, and managing robust data pipelines across batch, streaming, and real-time integration paradigms. The candidate will have 10+ years’ professional experience in Information Technology sector and at least 5 years of proven experience in Data Engineering Projects or Data Governance Teams operating in Large Corporations, preferable in the Telecommunication industry or the Finance Industry is highly preferable.

These pipelines will operate within Data Warehouse, Data Lake, and Data Lakehouse environments, leveraging technologies such as Teradata Vantage, Databricks, and Azure Cloud Data Services.

A strong command of Relational and Multidimensional Data Modelling is essential, along with advanced programming skills in Python, PySpark, and SQL. Experience with Geospatial data processing and Graph Databases is highly desirable.

The candidate should also demonstrate knowledge of core computer science principles, including Algorithms and Data Structures, Computer Architecture, Operating Systems, Programming Languages, Compilers and Interpreters, Theory of Computation, etc.

This role offers the opportunity to contribute to the evolution of a modern data ecosystem, enabling real-time insights and data-driven decision-making at scale.

Key Responsibilities

Design and implement reliable and efficient data pipelines on Vantage, Databricks and Azure Data Services.

Integrate and transform data from diverse sources including relational databases, cloud object storage, Event Stream Processing platforms, Web Services.

Develop and maintain scalable data architectures to support business intelligence, machine learning and Agentic AI initiatives.

Model complex datasets using both relational and multidimensional approaches (e.g., star/snowflake schemas).

Work with geospatial datasets and graph-based data structures to support advanced analytics use cases.

Collaborate with cross-functional teams including Data Scientists, Analysts, and Business Stakeholders.

Ensure data quality, governance, and security across all engineering processes.

Monitor and optimize performance of distributed systems and data workflows.

Contribute to the evolution of the company’s data strategy and platform architecture.

Education

Master’s or PhD in Computer Science, Engineering, Statistics, Physic, Mathematics.

Required Technical Skills

Data platforms: Databricks (Apache Spark), Azure Cloud Data Services (Fabric, Synapse, Functions, etc.), Teradata Vantage

Languages: Python (including PySpark), SQL, Unix Scripting

Relational databases: Teradata Vantage, SQL Server, PostgreSQL, MySQL

Geospatial data processing: GeoPandas, PostGIS, SQL Geospatial

Graph Databases: Neo4j, Cosmos DB with Gremlin

Data Modelling: Relational Modeling (3NF), Multidimensional Modeling (star/snowflake schema, slow changing dimension, etc.)

Soft Skills \& Leadership Competencies

Strong analytical thinking and problem-solving capabilities.

Excellent communication skills, with the ability to engage both technical and non-technical stakeholders.

Collaborative mindset and ability to work effectively in cross-functional teams.

High level of initiative, autonomy, and ability to manage multiple priorities.

Adaptability to fast-paced environments and emerging technologies.

Technical leadership and mentoring capabilities.

Strategic vision and business-oriented approach to data solutions.

Preferred Qualifications

Certifications in Azure, Databricks and Teradata.

Experience with CI/CD pipelines, Git, and DevOps practices.

Familiarity with data governance frameworks and GDPR compliance.

Competenze richieste

Se le può interessare , mi può allegare un suo cv alla mail : m.dangelo@dacomat.com esplicitando :

disponibilità in termini di tempistiche (eventuale preavviso)

aspettativa economica (ral annuale ,netto mensile desiderato / daily rate )

Il presente annuncio è rivolto a entrambi i sessi, ai sensi delle leggi 903/77 e 125/91, e a persone di tutte le età e tutte le nazionalità, ai sensi dei decreti legislativi 215/03 e 216/03.

Get job updates in your inbox

Subscribe to our newsletter and stay updated with the best job opportunities.