👨🏻‍💻 postech.work

Senior Data Pipeline Engineer - Freelance (80-100%, remote/Zurich/Berlin)

Decentriq • 🌐 Remote

Remote Posted 1 week, 1 day ago

Job Description

Decentriq is the rising leader in data-clean-room technology. With Decentriq, advertisers, retailers, and publishers securely collaborate on 1st-party data for optimal audience targeting and campaign measurement. Headquartered in Zürich, Decentriq is trusted by renowned institutions in the DACH market and beyond, such as RTL Ad Alliance, Publicis Media, and PostFinance.

Our analytics \& ML pipelines are the heartbeat of this platform. Built in Python and Apache Spark, they run in Databricks workspaces. We are looking for a

Senior Data Pipeline Engineer

freelancer

(≥ 80 %, start as soon as possible) for 6 months (possible conversion into FTE) to support our team during a crunch time driven by customer demand. The role can be fully remote (± 4 h CET) or based in our Zürich/Berlin office.

Would you like to help us make the advertising industry ready for the 1st-party era? Then we’d love to hear from you!

Tasks

Own, Design, Build \& Operate Data Pipelines

– Take responsibility for our Spark-based pipeline, from development through production and monitoring.

Advance our ML Models

– Improve and productionise models for AdTech use-cases such as lookalike modelling and demographics modeling.

AI-Powered Productivity

– Leverage LLM-based code assistants, design generators, and test-automation tools to move faster and raise the quality bar. Share your workflows with the team

Drive Continuous Improvement

– Profile, benchmark, and tune Spark workloads, introduce best practices in orchestration \& observability, and keep our tech stack future-proof.

Requirements

(Must have)

Bachelor/Master/PhD in Computer Science, Data Engineering, or a related field and

5+ years

of professional experience.

(Must have)

Expert-level Python and PySpark/Scala Spark experience

(Must have)

Proven track record building resilient, production-grade data pipelines with rigorous data-quality and validation checks.

(Must have)

Data-platform skills: operating Spark clusters, job schedulers, or orchestration frameworks (Airflow, Dagster, custom schedulers).

(Plus)

Working knowledge of ML lifecycle and model serving; familiarity with techniques for audience segmentation or look-a-like modelling is a big plus.

(Plus)

Rust proficiency (we use it for backend services and compute-heavy client-side modules).

Benefits

Competitive rates

Ownership instead of simply an executor

An amazing and fun team that is distributed all over Europe.

No need for a formal motivational letter. Just send your CV along with a few bullet points about why you're excited to work with us. We look forward to your application!

Get job updates in your inbox

Subscribe to our newsletter and stay updated with the best job opportunities.