This specialized role is paramount for designing, maintaining, and optimizing robust data pipelines and semantic models on our Azure-based Data Analytics Platform, leveraging Databricks and Microsoft Fabric. The ideal candidate combines strong technical proficiency in modern data engineering with the ability to translate finance and controlling business logic into governed, performant data models.
Experience with SAP FI/CO processes is preferred, as well as advanced skills in data modeling, Data Contracts, and cost/performance optimization. You will be instrumental in ensuring high data quality, governance, and availability for critical business intelligence and analytical dashboards. We are looking for a proactive, solution-oriented individual eager to contribute to a multidisciplinary, agile, and international environment.
- Design, build, and operate data pipelines on Azure Data Factory and Databricks (PySpark/SQL, Delta Lake) using Azure DevOps for CI/CD.
- Apply advanced data modeling techniques (dimensional/star, data vault, normalized models) and implement Medallion architecture (Bronze/Silver/Gold).
- Define and enforce Data Contracts: schemas, SLAs/SLOs, versioning, and validation gates.
- Optimize Databricks workloads for performance and cost (partitioning, Z ORDER, caching, Photon, autoscaling, cluster policies).
- Standardize delivery with Databricks Asset Bundles and implement observability (job metrics, audit logs).
- Ensure compliance with governance, security, and regulatory requirements via Unity Catalog and RBAC/ABAC policies.
- Embed data quality frameworks, automated tests, and monitoring for pipeline health, SLA breaches, and anomaly detection.
- Collaborate closely with Finance stakeholders and domain engineers to ensure KPI sign-off and business alignment.
- Contribute to technical documentation, participate in code reviews, and drive continuous improvement.
- (Preferred) Build semantic models in Microsoft Fabric/Power BI aligned with curated data and governed KPIs.
- (Preferred) Translate SAP FI/CO business logic (GL, AP/AR, allocations, exchange rates) into reconciled semantic models.
- Strong experience with Microsoft Azure (ADLS Gen2, Data Factory, Key Vault) and foundational networking/security.
- Hands-on expertise in Databricks: PySpark, SQL, Delta Lake, Unity Catalog, Asset Bundles; performance tuning and cost optimization.
- Advanced data modeling skills: dimensional/star, data vault, semantic layers; optimization for query performance.
- Proficiency in Python and SQL for data processing; modular code and unit testing.
- Experience with Azure DevOps (Repos, Pipelines, approvals) and CI/CD strategies with rollback procedures.
- Knowledge of Data Contracts: schema definition, SLAs/SLOs, versioning, compatibility policies.
- Familiarity with event-driven architectures and real-time data streaming.
- Experience working in Agile/Scrum environments.
- Fluent in English (written and spoken).
Preferred / Nice to Have
- SAP FI/CO domain knowledge (GL, AP/AR, Asset Accounting, Cost Center Accounting, Internal Orders, CO PA).
- Microsoft Fabric / Power BI: semantic modeling, dataset governance, KPI standardization.
- Infrastructure as Code (Terraform for Azure & Databricks).
- Data Quality & Anomaly Detection frameworks (DLT expectations, Great Expectations).
- Cost governance: tagging, dashboards, budgets/alerts.
- Advanced modeling patterns: slowly changing dimensions, snapshotting, late-arriving facts.
- Security & Compliance: data masking, tokenization, PII minimization.
B. Braun Medical, S.A. U. | Aitana Altuna