Skip to content

shannonlowder.com

Menu
  • About
  • Biml Interrogator Demo
  • Latest Posts
Menu

70-450: Calculating Table Size

Posted on November 3, 2011January 4, 2012 by slowder

In my previous three posts I covered different RAID solutions, and when I use each type of RAID. That’s just the first part of the first section of the 70-450.

Designing a SQL Server Instance and a Database Solution (14 percent)

  • Design for CPU, memory, and storage capacity requirements.
    • This objective may include but is not limited to: RAID, calculating table size, IO throughput, transaction per second, data compression, non-uniform memory access (NUMA), tempdb capacity

Tape Measure-at-244-by-G.E.Sattler

Let’s move on to the next topic, calculating table size. I’ll admit, when I get questions about table size it’s usually how large is the table now, and how has it grown over time. That way we can make the call on whether or not we need to move the database to a new disk with more space, or add a data file to the database, or simply allow the database more room to grow.

When I did some digging into the MSDN, it looks like they want you to be able to calculate the table sizes based on table designs. So after you’ve decided what columns you’ll need, what data types those columns will be, whether or not those columns are nullable, and if you’ll use a heap or a clustered index, and if you’ll include any non-clustered indexes.

After some googling, I found a book that tries to cover this topic: SQL Server 2008 Administration: Real-World Skills for MCITP Certification. The explanation in the book is a great mass of text explaining how to do the calculations.

Turning to the MSDN, I found A little more help here. Again, it’s still a mass of text(s). The only way I could think to make this estimation make more sense was to build out an excel spreadsheet.

The problem is you’re not going to be able to take that spreadsheet in with you on an exam. I suggest you build your own spreadsheet based on the calculations give on the three pages: Estimating the Size of a Heap, , and Estimating the Size of a Nonclustered Index.

I have to admit this estimation is a bit detailed compared to what I’ve had to do in the past. Most of the time, I have to anticipate the volume of rows a table will hold, multiply that by the row size, then double that number to give myself room to grow.

The extra details I learned while exploring these calculations include:

  • null bitmaps – is a part of a row that if a column is null or not. It’s size is 2 + ((Number of columns + 7) / 8).
  • Microsoft suggests estimating the size of a variable length field as 2 + (Number of variable Columns * 2) + max variable length. I usually just added up the max length of all columns, rather than doing this calculation, I wanted more buffer in my estimates.
  • when calculating row size in bytes, add 4 for the row header.
  • When calculating clustered index space requirements, don’t forget to include information about the index you’re making (remember: clustered indexes are the physical order of the table, you have to store that information somewhere!)
  • Calculating non-clustered indexes is similar to calculating a clustered index, but you only count included and indexed columns.

Studying

This material is worth reviewing, but only for the exam. In production, you’d use the excel spreadsheet, or another tool to actually build out your size estimates.

Next time, We’ll hit IO throughput. You’ll see me referencing the RAID configurations I told you about last time, as well as adding filegroups, files and other slight of hand in order to maximize throughput. Until then, if you have any questions, send them in. I’m here to help!

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Recent Posts

  • A New File Interrogator
  • Using Generative AI in Data Engineering
  • Getting started with Microsoft Fabric
  • Docker-based Spark
  • Network Infrastructure Updates

Recent Comments

  1. slowder on Data Engineering for Databricks
  2. Alex Ott on Data Engineering for Databricks

Archives

  • July 2023
  • June 2023
  • March 2023
  • February 2023
  • January 2023
  • December 2022
  • November 2022
  • October 2022
  • October 2018
  • August 2018
  • May 2018
  • February 2018
  • January 2018
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • June 2017
  • March 2017
  • February 2014
  • January 2014
  • December 2013
  • November 2013
  • October 2013
  • August 2013
  • July 2013
  • June 2013
  • February 2013
  • January 2013
  • August 2012
  • June 2012
  • May 2012
  • April 2012
  • March 2012
  • February 2012
  • January 2012
  • December 2011
  • November 2011
  • October 2011
  • September 2011
  • August 2011
  • July 2011
  • June 2011
  • May 2011
  • April 2011
  • March 2011
  • February 2011
  • January 2011
  • December 2010
  • November 2010
  • October 2010
  • September 2010
  • August 2010
  • July 2010
  • June 2010
  • May 2010
  • April 2010
  • March 2010
  • January 2010
  • December 2009
  • November 2009
  • October 2009
  • September 2009
  • August 2009
  • July 2009
  • June 2009
  • May 2009
  • April 2009
  • March 2009
  • February 2009
  • January 2009
  • December 2008
  • November 2008
  • October 2008
  • September 2008
  • August 2008
  • July 2008
  • June 2008
  • May 2008
  • April 2008
  • March 2008
  • February 2008
  • January 2008
  • November 2007
  • October 2007
  • September 2007
  • August 2007
  • July 2007
  • June 2007
  • May 2007
  • April 2007
  • March 2007
  • February 2007
  • January 2007
  • December 2006
  • November 2006
  • October 2006
  • September 2006
  • August 2006
  • July 2006
  • June 2006
  • May 2006
  • April 2006
  • March 2006
  • February 2006
  • January 2006
  • December 2005
  • November 2005
  • October 2005
  • September 2005
  • August 2005
  • July 2005
  • June 2005
  • May 2005
  • April 2005
  • March 2005
  • February 2005
  • January 2005
  • November 2004
  • September 2004
  • August 2004
  • July 2004
  • April 2004
  • March 2004
  • June 2002

Categories

  • Career Development
  • Data Engineering
  • Data Science
  • Infrastructure
  • Microsoft SQL
  • Modern Data Estate
  • Personal
  • Random Technology
  • uncategorized
© 2025 shannonlowder.com | Powered by Minimalist Blog WordPress Theme