Automated Jira Ticket Report Generation Per School Using Fabric & Power BI

It would be helpful for school admins or principals to see ticket activity per school per week. Seeing data like how many tickets in total, ticket number per user, tickets that were resolved vs opened would be helpful for admins to see. To do this, we should organize the data based on school first, so […]

Automating Custom Field Bulk Edits in Jira Post Migration with Fabric

We have automated around 100k tickets over to Jira and to clean the Jira production data, some fields need to be altered. By default, some tickets were assigned to Jin as a default user, but this is problematic for reporting and filtering purposes in the future. As such, we needed to change around 60k tickets […]

Automating Knowledge Base/ Documentation Transfer From Onenote to Confluence using Fabric

The general logic of executing the Onenote data ingest: Onenote data ingest PySpark code: import requests from pyspark.sql import SparkSession from pyspark.sql.types import StructType, StructField, StringType # Initialize Spark session spark = SparkSession.builder.appName(“OneNoteDataPipeline”).getOrCreate() # Authentication def get_access_token():     url = “https://login.microsoftonline.com/982d56ce-f6e7-4334-a1c4-5d6779c789a6/oauth2/v2.0/token”     payload = {         “grant_type”: “client_credentials”,     […]

Leveraging PySpark to Automate Jira Ticket Creation

Trying to create a pyspark script in Microsoft Fabric to automate JIRA API ticket creation (post) calls against a lakehouse delta table. To do so, I have to educate myself on the pyspark syntax and framework. I will be doing so in this blog post. To reference a lakehouse delta table: Result: This was able […]

Automating Ticket Migration for Jira

I have been tasked with automating the migration of over 100k SolarWinds Helpdesk tickets over to Jira Service Management as we are preparing for the helpdesk transition. Plan: My current plan is to create an ETL on Microsoft Fabric that ingests CSV ticket data from Solarwinds and manipulate field values to fit our specific Jira […]

Microsoft Fabric & Spark SQL

The most common way to work with data in delta tables in Spark is to use Spark SQL. Lets say we have a table in OneLake called products: In the connected Spark SQL, we can insert data like so: When you want to work with delta files rather than catalog tables, it may be simpler […]

Microsoft Fabric: Optimize Performance by Partitioning Delta Tables

In OneLake, we are able to partition data so that performance enhancements could be made through data skipping. Consider a situation where large amounts of sales data are being stored. You could partition sales data by year. The partitions are stored in subfolders named “year=2021”, “year=2022”, etc. If you only want to report on sales […]

Microsoft Fabric: Optimizing Database Table Architecture

Very simply, we are able to optimize lakehouse data table architecture. In the case where Spark is used, Parquet files are immutable and as such, we end up storing a lot of small files. The Optimize function allows us to reduce the number of files written as it just collates them into larger files. Within […]