Back to job search

PowerBI Engineer

  • Location:

    Glasgow

  • Sector:

    Data & AI

  • Job type:

    Permanent

  • Salary:

    £50000 - £62000 per annum

  • Contact:

    Eve Fraser

  • Contact email:

    efraser@headresourcing.com

  • Job ref:

    BBBH37906_1763377104

  • Published:

    about 4 hours ago

  • Expiry date:

    2025-12-17

  • Startdate:

    ASAP

? Power BI Report Engineer (Azure / Databricks)

Glasgow | 3-4 days onsite | Exclusive Opportunity with a Leading UK Consumer Brand


Are you a Power BI specialist who loves clean, governed data and high-performance semantic models?
Do you want to work with a business that's rebuilding its entire BI estate the right way-proper Lakehouse architecture, curated Gold tables, PBIP, Git, and end-to-end governance?
If so, this is one of the most modern, forward-thinking Power BI engineering roles in Scotland.
Our Glasgow-based client is transforming its reporting platform using Azure + Databricks, with Power BI sitting on top of a fully curated Gold Layer. They develop everything using PBIP + Git + Tabular Editor 3, and semantic modelling is treated as a first-class engineering discipline.
This is your chance to own the creation of high-quality datasets and dashboards used across Operations, Finance, Sales, Logistics and Customer Care-turning trusted Lakehouse data into insights the business relies on every day.


? Why This Role Exists


To turn clean, curated Gold Lakehouse data into trusted, enterprise-grade Power BI insights.
You'll own semantic modelling, dataset optimisation, governance and best-practice delivery across a modern BI ecosystem.


? What You'll Do

Semantic Modelling with PBIP + Git


  • Build and maintain enterprise PBIP datasets fully version-controlled in Git.

  • Use Tabular Editor 3 for DAX, metadata modelling, calc groups and object governance.

  • Manage branching, pull requests and releases via Azure DevOps.

Lakehouse-Aligned Reporting (Gold Layer Only)


  • Develop semantic models exclusively on top of curated Gold Databricks tables.

  • Work closely with Data Engineering on schema design and contract-first modelling.

  • Maintain consistent dimensional modelling aligned to the enterprise Bus Matrix.

High-Performance Power BI Engineering


  • Optimise performance: aggregations, composite models, incremental refresh, DQ/Import strategy.

  • Tune Databricks SQL Warehouse queries for speed and cost efficiency.

  • Monitor PPU capacity performance, refresh reliability and dataset health.

Governance, Security & Standards


  • Implement RLS/OLS, naming conventions, KPI definitions and calc groups.

  • Apply dataset certification, endorsements and governance metadata.

  • Align semantic models with lineage and security policies across the Azure/Databricks estate.

Lifecycle, Release & Best Practice Delivery


  • Use Power BI Deployment Pipelines for Dev ? UAT ? Prod releases.

  • Enforce semantic CI/CD patterns with PBIP + Git + Tabular Editor.

  • Build reusable, certified datasets and dataflows enabling scalable self-service BI.

Adoption, UX & Collaboration


  • Design intuitive dashboards with consistent UX across multiple business functions.

  • Support BI adoption through training, documentation and best-practice guidance.

  • Use telemetry to track usage, performance and improve user experience.

? What We're Looking For

Required Certifications


To meet BI engineering standards, candidates must hold:


  • PL-300: Power BI Data Analyst Associate

  • DP-600: Fabric Analytics Engineer Associate

Skills & Experience


  • 3-5+ years building enterprise Power BI datasets and dashboards.

  • Strong DAX and semantic modelling expertise (calc groups, conformed dimensions, role-playing dimensions).

  • Strong SQL skills; comfortable working with Databricks Gold-layer tables.

  • Proven ability to optimise dataset performance (aggregations, incremental refresh, DQ/Import).

  • Experience working with Git-based modelling workflows and PR reviews via Tabular Editor.

  • Excellent design intuition-clean layouts, drill paths, and KPI logic.

Nice to Have


  • Python for automation or ad-hoc prep; PySpark familiarity.

  • Understanding of Lakehouse patterns, Delta Lake, metadata-driven pipelines.

  • Unity Catalog / Purview experience for lineage and governance.

  • RLS/OLS implementation experience.

If this job isn't the one for you, then don't worry we have lots more opportunities available!