You are opening our Spanish language website. You can keep reading or switch to other languages.

Senior Data Engineer, Clinical Data Platform

  • Almatý
  • Astaná
  • Belgrado
  • Breslavia
  • Cluj-Napoca
  • Cracovia
  • Ereván
  • Lárnaca
  • Lodz
  • Lublinie
  • Remote.Bulgaria
  • Remote.Georgia
  • Remote.Kazakhstan
  • Remote.Poland
  • Riga
  • Sofia
  • Tiflis
  • Varna
  • Varsovia
Equipo mediano (10-20 personas)Hot vacancy

Si has recibido esta oferta laboral de parte de nuestros reclutadores, te pedimos que leas nuestro Aviso de Privacidad.

Project overview

You will work on a platform that processes clinical and real-world data (EHRs, labs, registries, trial data) and powers analytics, reporting, and data products for a healthcare / clinical research client.

Position overview

We are looking for a Senior Data Engineer to build and operate a clinical data platform on Databricks, with a strong focus on robust data pipelines, data models, and data quality.

Technology stack

The platform is built on Databricks (Spark, Delta Lake) and includes reusable pipelines, a shared data model, and automated data quality checks.

Responsibilities

  • Design, build, and maintain end-to-end Databricks data pipelines (ingestion, transformation, publishing) for production use
  • Work with data models (staging, curated, canonical, or dimensional) and help evolve them together with architects and analysts
  • Embed data quality and data governance rules into all pipelines (checks, validation, monitoring, alerting)
  • Optimize Databricks jobs for performance and cost (cluster configuration, partitioning, caching, file layout)
  • Collaborate with data architects, analysts, and domain experts to clarify requirements and refine technical solutions

Requirements

  • 5+ years of experience in data engineering, DWH, or big data, including production data pipelines
  • Strong hands-on experience with Databricks: Spark (PySpark/Scala), Delta Lake, Databricks Jobs / Workflows
  • Proven experience designing and operating end-to-end pipelines on Databricks for batch or near-real-time data
  • Experience with data pipelines and CI/CD for data
  • Practical experience with data modeling (layered models, canonical or dimensional models) for analytics and reporting
  • Experience embedding data quality and data governance rules into pipelines (schema checks, business rules, SLOs, monitoring)
  • Good communication skills, upper-intermediate or higher English proficiency, and the ability to work closely with stakeholders in distributed teams and communicate directly with clients

Nice to have

  • Experience designing and delivering PoC solutions on Databricks to quickly validate ideas using real data
  • Experience with ontologies or a semantic layer (business concepts, metrics, mappings) on top of analytical data

Looking for Similar Opportunities?

Try AI chatbots with our ready-made prompt to discover similar roles that match your skills and interests.
Image

We offer

Image

Trabajo remoto

Ofrecemos una gran flexibilidad para trabajar desde distintas ciudades y países

Image

Días off para descansar

Todos los colegas cuentan con días off para viajar, descansar y pasar tiempo con sus seres queridos

Image

Feriados nacionales

Según el calendario oficial de cada país

Image

Días off por maternidad y paternidad

Todos los colegas disfrutan de días off para compartir con su bebé

Image

Certificaciones pagas

Impulsamos el desarrollo profesional y certificación de nuestros colegas

Image

Plataforma de e-learning interna

Acceso ilimitado a cursos y entrenamientos

Image

Clases de idiomas

Clases de inglés virtuales con profesoras altamente calificadas

Image

Comunidades profesionales

Todos los colegas pueden participar de comunidades profesionales internacionales y regionales, en base a sus intereses

El paquete de beneficios puede variar según la región y el tipo de contrato.
Más buscadas
1 of 1