Make Your Resume Now

Data Warehouse and Reporting Developeer

Posted December 11, 2025
Full-time Mid-Senior level

Job Overview

At Uni Systems, we are working towards turning digital visions into reality. We are continuously growing and we are looking for a Data Warehouse and Reporting Developer to join our UniQue team.

What will you be doing in this role?

  • Develop, deploy, and maintain scalable and incremental data pipelines from REST APIs and databases using Python, PySpark, Azure Synapse, Knime, SQL, and ETL tools to ingest, transform, and prepare data.
  • Process and transform complex JSON and GIS data into structured datasets optimized for analysis and reporting. This includes parsing, transforming, and validating JSON data to ensure data quality and consistency.
  • Load, organize, and manage data in Azure Data Lake Storage and Microsoft Fabric OneLake, ensuring accessibility, performance, and efficient storage using lakehouse and Delta Lake patterns.
  • Document ETL processes, metadata definitions, data lineage, and technical specifications to ensure transparency and reusability.
  • Collaborate with data analysts, BI developers, and business stakeholders to understand data requirements and deliver reliable, well-documented datasets aligned with organizational needs.
  • Implement data quality checks, logging, monitoring, and automated incremental load mechanisms within data pipelines to support maintainability, observability, and troubleshooting

Requirements

What will you be bringing to the team?

  • Master's Degree and 12 years of experience or Bachelor's Degree and 16 years of experience in IT.
  • Hold Microsoft Azure Data Engineer Associate certificate.
  • At least 5 years of experience in Azure Data Lake Storage, Microsoft Fabric OneLake, and Oracle databases.
  • Minimum 5 years of experience in developing data pipelines from REST APIs and on integration (such as Azure Synapse, PySpark, Microsoft Fabric, Python, SQL, KNIME).
  • No less than 5 years of experience in processing JSON and GIS data.
  • Excellent knowledge of data engineering tools Azure Synapse Analytics, Microsoft Fabric, PySpark and Python.
  • Experience designing incremental loads, CDC processes, and automated schema evolution
  • Ability to implement robust data quality checks, logging, and monitoring in ETL processes.
  • Ability to document ETL workflows, metadata, and technical specifications clearly and consistently.
  • Familiarity with DevOps and version control best practices. Experience with CI/CD pipelines.
  • Experience working in an Agile and Scrum framework.
  • Proficiency in English language at a C1/C2 level.
  • Proficiency in French language is considered an advantage.

Ready to Apply?

Take the next step in your career journey

Stand out with a professional resume tailored for this role

Build Your Resume – It’s Free!