Make Your Resume Now

Azure Data Engineer

Posted February 20, 2026
Full-time Mid-Senior Level

Job Overview

We are seeking an experienced Azure Data Engineer to join our enterprise data engineering team. This role is focused on building and maintaining modern, scalable data pipelines across our data ecosystem — including lakehouses, data warehouses, data marts, and operational data stores — while supporting the migration of legacy ETL solutions to Microsoft Fabric and Azure.

Key Responsibilities:

Data Pipeline Development 

  • Design and build ETL/ELT pipelines using Azure Data Factory, Microsoft Fabric Data Pipelines, Databricks, and Fabric Notebooks
  • Implement medallion architecture (Bronze/Silver/Gold) in Fabric Lakehouse environments
  • Develop transformation logic using T-SQL, Spark SQL, PySpark, and Dataflows Gen2
  • Build and maintain dimensional models (star/snowflake schema) and Data Vault models
  • Implement incremental loading patterns using CDC, watermarking, and delta detection
  • Create reusable pipeline components, templates, and parameterized frameworks
  • Optimize pipeline performance through partitioning, parallelization, and query tuning

Legacy-to-Fabric Migration

  • Convert legacy ETL mappings, workflows, and scheduling logic to Microsoft Fabric/ADF equivalents
  • Recreate parameter files, session configurations, and orchestration patterns in Fabric
  • Execute unit testing and data reconciliation to validate migrated pipelines produce identical results
  • Document conversion patterns, technical decisions, and issue resolutions
  • Support parallel runs and cutover validation

Data Quality & Testing 

  • Build data quality checks and validation frameworks embedded within pipelines
  • Develop automated testing strategies (unit, integration, regression) for data pipelines
  • Create monitoring dashboards and alerting for pipeline failures and data anomalies
  • Perform source-to-target reconciliation for both BAU and migration workloads

Platform Operations & Collaboration 

  • Monitor, troubleshoot, and optimize production pipelines
  • Implement logging, error handling, and retry mechanisms
  • Support CI/CD pipelines for data solutions using Azure DevOps and Git
  • Manage environment promotions (DEV → QA → PROD) and participate in on-call rotation
  • Implement security best practices: RBAC, encryption, data masking, workspace security
  • Collaborate with Data Architects, Business Analysts, DevOps, and BI teams
  • Maintain technical documentation: pipeline specs, data dictionaries, and runbooks

Technical Skills:

Microsoft Fabric & Azure

  • Microsoft Fabric — Lakehouse, Data Warehouse, Data Pipelines, Dataflows Gen2, Notebooks
  • Azure Data Factory v2 — pipelines, linked services, integration runtimes, triggers
  • Azure Synapse Analytics — Dedicated SQL Pools, Serverless SQL, Spark Pools
  • Azure Data Lake Storage Gen2, OneLake, Shortcuts, and Direct Lake mode

SQL & Programming

  • Expert-level T-SQL — stored procedures, complex queries, performance tuning
  • Python for data processing and automation
  • PySpark for large-scale data transformations
  • Familiarity with JSON, XML, and REST APIs

Informatica Platform

  • Development experience with Informatica PowerCenter (Designer, Workflow Manager, Workflow Monitor)

Data Platforms & Formats

  • Delta Lake format and Delta table operations
  • Apache Spark architecture and optimization
  • Data partitioning strategies and performance tuning
  • Parquet and Avro file formats
  • Dimensional modeling and Data Vault concepts

DevOps & Governance

  • Git version control and Azure DevOps (Repos, Pipelines)
  • CI/CD implementation for data solutions
  • Fabric workspace deployment pipelines
  • Data lineage, metadata management, and data cataloging
  • Security best practices — RBAC, encryption, masking
  • Awareness of compliance standards (GDPR, HIPAA, SOC2)

Ready to Apply?

Take the next step in your career journey

Stand out with a professional resume tailored for this role

Build Your Resume – It’s Free!