Comparing Fabric Capacity Consumption – Notebook vs Warehouse SQL
I saw that there was an update where it is now possible to use the Microsoft Fabric Warehouse to copy data directly from OneLake into the Warehouse. This got me thinking, which would consume more capacity to get the data into the Warehouse table. As well as which one would be faster. To do this I am going to be…
How to use the Tabular Object Model using Semantic Link Labs in a Microsoft Fabric Notebook
In this blog post I am going to show you how to use the powerful Semantic Link Labs library for Tabular Object Model (TOM) for semantic model manipulation. The goal of this blog post is to give you an understanding of how to connect using TOM, then based on the documentation use one of the functions. Don’t get me wrong…
Using a Python Notebook Loop through a data frame and write once to a Lakehouse Table
In this blog post I am going to explain how to loop through a data frame to query data and write once to a Lakehouse table. The example I will use is to loop through a list of dates which I get from my date table, then query an API, append to an existing data frame and finally write once…
Using a Python Notebook to Query Multiple Lakehouse Tables into a Data frame
In this blog post I am going to explain how to query multiple Lakehouse tables into a data frame. The example I am going to use is when you want to load new data into your staging tables, but you need to know the max date from your previous data load. I am going to build on my previous blog…
Using a Python Notebook using Semantic Link Labs to write a DAX Query output to a Lakehouse Table
In this blog post I am going to explain how to use a Python Notebook using the Semantic Link module, to run a DAX query and write the output to a Lakehouse table. I will show you how to install a Python library and then use it within my python notebook. If you are interested in more details on why…
How to use SQL Code to write to Delta Tables in a Python Notebook
In this blog post I am going to explain how to use SQL Code to write to Delta Tables in a Microsoft Fabric Lakehouse. If you are interested in more details on why I am using a Python Notebook please read this blog post: What are the advantages of using Python Notebook vs a Spark Notebook in Microsoft Fabric –…
How to use SQL Code to read from Delta Tables in a Python Notebook
I come from a TSQL background, so using SQL makes it easy for me to work with data. There are multiple ways to use SQL in a PySpark notebook, and when I started using a Python notebook it was not so straightforward. In this blog post I will show you how I use SQL Code. As mentioned previously I am…
What are the advantages of using Python Notebook vs a Spark Notebook in Microsoft Fabric
I would like to start off because I am by no means a Python expert. I have been learning Python for about 2 years 😊 This is the first blog post in a series of blog posts where I dive into how to use Python notebooks instead of Spark notebooks. For example, I will show you how to run a…
How to get all your Fabric Items using Python Only Notebook
This blog below explains how to use a Python only notebook to get all the Fabric items using the Fabric REST API. NOTE: At the time of this blog post Feb 2025, Dataflow Gen2 is not included in the Fabric items, I am sure it will be there in the future. NOTE II: This only gets the Fabric Items, which…
Getting Notified when a Microsoft Fabric Notebook fails
What I have found is that when I created a pipeline in Microsoft Fabric that uses a notebook, when there is an error with the notebook, I do not get an alert that the notebook has failed. This has happened to me in the past and I have found this pattern below to work consistently to notify me of errors….