Skip to content

shannonlowder.com

Menu
  • About
  • Biml Interrogator Demo
  • Latest Posts
Menu

Metadata Model Update

Posted on May 16, 2018November 14, 2022 by slowder

As I began learning Biml,  I developed my original metadata model to help automate as much of my BI development as I could.  This model still works today, but as I work with more file based solutions in Azure Data Lakes, and some “Big Data” solutions, I’m discovering it’s limitations. Today I’d like to talk through one possible solution to this problem: abstraction.  Before we get into that, I’d like to revisit my previous model in case you haven’t seen it before.

Current State

My model follows a hierarchy that will be familiar to anyone who’s spent time with Microsoft SQL Server.  It all begins with Connections.  You can have connections to database servers, files, or even an API endpoint.  I have a connection type attribute on the connection that let’s me know what I can and can’t do with the connection.  That’s important for choosing how to deal with a file versus dealing with a database.

When we move down a layer to Database, things have to be made to fit the model.  When we’re talking about files, they don’t have a concept of database.  They also don’t have a concept of Schema.  In many of my solutions, I’ll use Database and Schema to refer to folders in a file path.  A database will refer to one folder, schema will refer to a subfolder of that “database” folder. Then inside that “schema” folder you’ll find folders where every file in a given folder is of a given layout.  All these files are to be processed into a given table in SQL Server.

But what happens when you don’t have a three folder deep path?

What if your connection is an API end point and each request is a different “table”?

I keep finding more and more situations where my metadata model is too limiting to my solutions. Fortunately, my concept of Table and Column fit nearly every situation I’ve come up against so far.  Even if the name “Table” actually refers to a view, stored procedure output, or file’s contents.

New Solution

I struggled with how to extend my current model for a while.  I’d understood object oriented programming concepts like objects, inheritance, and polymorphism, but they’d just become words developers used to explain how their code was so awesome. As I grew into more of a developer by building out my BI tool belt, these words started to make sense.  In fact, they gave me a new idea on how to solve the metadata model problem.

I kept my Connection records like they were, but I effectively combined my concepts of Database, Schema, and Table into a single concept: Object. An object in this new model has a connection, type, and optionally a parent object.

If I want to model a database, schema, and table, I can still do that!  I just insert a row for the database into the object table, with a reference to it’s connection and type “database”. Then I grab SCOPE_IDENTITY() for that insert statement.  That gives me the parent object ID I’ll use for the next insert: schema.  I insert the schema name, connection ID, and the database’s identity value, along with a type “schema”.  I grab the next SCOPE_IDENTITY() generated so I’ll have it for my table insert.  Finally, I insert the table’s name, connection ID, and the schema’s ID, along with a type of “table”.

Then, when I go to model columns, I do that like I would for the previous model.

Now, when I want to model a file, I can model a connection to a single file with just a Connection record and a Object record.  I could also model a complex folder structure full of files with multiple Object records of type “folder”.

I could also model a connection and data object for each API call I want to mine for information.

No more forcing a data model into an arbitrary structure.  I now have flexibility!

Full Disclosure

I’m actually cutting over all my previous solutions to this new model to see if I can break it with existing examples.  My hope is I can move my old solutions onto this new metadata model without breaking anything.  I know I’ll have to rebuild my code that builds SSIS packages to get the data from the new model. I’m also moving all my new projects to this model too, since many more projects are file based now than they were before.  If I find any examples that break this model, I’ll share my findings here.

If you have any thoughts on this model change, I’d love to hear them!

Until next time, keep on automating!

 

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Recent Posts

  • A New File Interrogator
  • Using Generative AI in Data Engineering
  • Getting started with Microsoft Fabric
  • Docker-based Spark
  • Network Infrastructure Updates

Recent Comments

  1. slowder on Data Engineering for Databricks
  2. Alex Ott on Data Engineering for Databricks

Archives

  • July 2023
  • June 2023
  • March 2023
  • February 2023
  • January 2023
  • December 2022
  • November 2022
  • October 2022
  • October 2018
  • August 2018
  • May 2018
  • February 2018
  • January 2018
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • June 2017
  • March 2017
  • February 2014
  • January 2014
  • December 2013
  • November 2013
  • October 2013
  • August 2013
  • July 2013
  • June 2013
  • February 2013
  • January 2013
  • August 2012
  • June 2012
  • May 2012
  • April 2012
  • March 2012
  • February 2012
  • January 2012
  • December 2011
  • November 2011
  • October 2011
  • September 2011
  • August 2011
  • July 2011
  • June 2011
  • May 2011
  • April 2011
  • March 2011
  • February 2011
  • January 2011
  • December 2010
  • November 2010
  • October 2010
  • September 2010
  • August 2010
  • July 2010
  • June 2010
  • May 2010
  • April 2010
  • March 2010
  • January 2010
  • December 2009
  • November 2009
  • October 2009
  • September 2009
  • August 2009
  • July 2009
  • June 2009
  • May 2009
  • April 2009
  • March 2009
  • February 2009
  • January 2009
  • December 2008
  • November 2008
  • October 2008
  • September 2008
  • August 2008
  • July 2008
  • June 2008
  • May 2008
  • April 2008
  • March 2008
  • February 2008
  • January 2008
  • November 2007
  • October 2007
  • September 2007
  • August 2007
  • July 2007
  • June 2007
  • May 2007
  • April 2007
  • March 2007
  • February 2007
  • January 2007
  • December 2006
  • November 2006
  • October 2006
  • September 2006
  • August 2006
  • July 2006
  • June 2006
  • May 2006
  • April 2006
  • March 2006
  • February 2006
  • January 2006
  • December 2005
  • November 2005
  • October 2005
  • September 2005
  • August 2005
  • July 2005
  • June 2005
  • May 2005
  • April 2005
  • March 2005
  • February 2005
  • January 2005
  • November 2004
  • September 2004
  • August 2004
  • July 2004
  • April 2004
  • March 2004
  • June 2002

Categories

  • Career Development
  • Data Engineering
  • Data Science
  • Infrastructure
  • Microsoft SQL
  • Modern Data Estate
  • Personal
  • Random Technology
  • uncategorized
© 2025 shannonlowder.com | Powered by Minimalist Blog WordPress Theme