Automating your Power BI dataset refresh can save time and ensure your reports always stay up to date. In this post, we’ll walk through how to trigger a Power BI report refresh directly from Azure Data Factory (ADF) using an App Registration in Azure Entra ID.
5 Career Skills Every Aspiring Data Engineer Needs
Over the past ten years in data, I’ve been fortunate to collaborate with, learn from, and help grow some incredibly talented engineering teams. One thing has become clear along the way: Tools come and go, but the core skills that make a great data engineer remain timeless. In today’s cloud-first, real-time, business-aligned world, here are the five skills I believe every aspiring data engineer must master.
What Master Data Management (MDM) Really Means in 2025
At its core, MDM is the practice of defining, managing, and governing the most critical data entities such as customers, products, suppliers, locations, employees so that the entire organization agrees on a single version of the truth.
How to Extract HubSpot Data to Azure Data Lake Using Azure Data Factory
In today’s data-driven world, integrating marketing and sales data from platforms like HubSpot into a centralized storage system such as Azure Data Lake is critical for advanced analytics and reporting. Azure Data Factory (ADF) makes it simple to automate this process without writing any code. In this guide, we'll walk through the exact steps needed to pull data from HubSpot and store it in Azure Data Lake Storage Gen2 in JSON format.
How to Scale Up and Scale Down Dedicated SQL pool (SQL DW) using Azure Data factory.
Scaling up and scaling down your Azure Dedicated SQL Pool helps optimize both performance and costs.
How to Load Parquet Files from Azure Data Lake to Data Warehouse
By following these steps, you’ll be able to extract, transform, and load (ETL) your Parquet data into a structured data warehouse environment, enabling better analytics and reporting.
How to Copy Data from JSON to Parquet in Azure Data Lake
In this step-by-step guide, we’ll go through the exact process of creating Linked Services, defining datasets, and setting up a Copy Activity to seamlessly transfer your JSON data to Parquet format.