This article showcases how to take advantage of a highly distributed framework provided by spark engine, to load data into a Clustered Columnstore Index of a relational database like SQL Server or Azure SQL Database, by carefully partitioning the data before insertion.
(image)
I'm preparing a series of post and samples on how to properly load data into Azure SQL using Azure Databricks / Apache Spark that I will start to publish very soon, but I realized today that there is a pre-requisite that in many cases, especially by developers new to the data space, is overlooked: good table design.
Wait! If you'...