I talk a lot about compression. I’ve blogged a pretty decent amount on it as well. One of the things that often confuses people is what can and cannot be compressed. There is a list of data types that can be Row Compressed. That list is different between each SQL Version. Page compression on the other hand works at the binary level, it is data type agnostic.
The big determining factor is what type of Allocation Unit your data is stored on.
“Balls,” you say “What’s an Allocation Unit?”
An Allocation unit is the structure behind the structure. Think of real estate for a second. Buildings and property are zoned in a city or a town. One section is for businesses, another is zoned for residential, one may be zoned for the government. In SQL Server we have 3 different zones IN_ROW_DATA, ROW_OVERFLOW_DATA, and LOB_DATA.
Instead of being sized just for type, your size matters just as much. If you are a regular every day Integer or Character field you live in IN_ROW_DATA. You are LOB_DATA if you are a VARBINARY(MAX) that contains a 500 MB picture file. ROW_OVERFLOW_DATA are variable length fields that start off on IN_ROW_DATA pages, but if that data grows large enough that it cannot fit on an 8 KB IN_ROW_DATA page then it gets popped off the IN_ROW_DATA Page and lands on the ROW_OVERFLOW_DATA Page.
The data types in SQL that have a (MAX) designation, XML, or certain CLR types start off on IN_ROW_DATA pages. They get moved off if the size grows.
HOW IN THE WIDE WIDE WORLD OF SPORTS
So how in the wide wide world of sports does this apply to Data Compression? If your data is on an IN_ROW_DATA page it could be compressed. Row compression still only applies to the data types that are listed per version, see row compression here at MSDN.
Page Compression only requires matching binary patterns, as long as it is IN_ROW_DATA pages we are good to go. You can use this script to run against your database to get the Allocation Unit makeup of your tables and indexes.
OBJECT_NAME(sp.object_id) AS [ObjectName]
,si.name AS IndexName
,sps.in_row_data_page_count as In_Row
,sps.row_overflow_used_page_count AS Row_Over_Flow
,sps.lob_reserved_page_count AS LOB_Data
JOIN sys.partitions sp
JOIN sys.indexes si
ON sp.index_id=si.index_id AND sp.object_id = si.object_id
order by sps.in_row_data_page_count desc
The higher the IN_ROW_DATA page count the more likely you have a candidate for compression.
ON TO THE MAIN EVENT
We’ve laid the ground work now on to the main event. First we’ll create our database and our table and insert some data. I’ve got two Varchar(Max) fields, we’ll put 2012 characters in each.
Select our demo database
if exists(select name from sys.databases where name='demoInternals')
alter database demoInternals set single_user with rollback immediate
drop database demoInternals
Create Database demoInternals
Create our table
IF EXISTS(SELECT name FROM sys.tables WHERE name='vmaxTest')
DROP TABLE dbo.vmaxTest
CREATE TABLE vmaxTest(myid int identity(1,1)
, mydata varchar(max) default 'a'
,mydata2 varchar(max) default 'b'
,CONSTRAINT pk_vmaxtest1 PRIMARY KEY CLUSTERED (myid))
Insert 5000 rows
DECLARE @i INT
INSERT INTO vmaxTest(mydata, mydata2)
VALUES(replicate('a',2012)+cast(@i AS VARCHAR(5)), replicate('b', 2012)+cast(@i AS VARCHAR(5)))
If you use our script from earlier then you can see we have 4950 IN_ROW_DATA Pages.
Now let’s update one of our Varchar(max) fields to 8000 characters so that we push it off of IN_ROW_DATA and over to LOB_DATA Pages. Run our script again to get our counts.
Now we'll update just the b values
to force them into row_overflow data
We certainly have some fragmentation, but we’ve added 5009 LOB_DATA pages to the mix. Now let’s apply Page Compression and use our script again to see the results.
Rebuild our table with
ALTER TABLE dbo.vmaxtest
As you can see the IN_ROW_DATA Compressed, the LOB_DATA didn’t. Another way that knowing thy data can help you understand what you can and should compress.