We use cookies and other similar technologies (Cookies) to enhance your experience and to provide you with relevant content and ads. By using our website, you are agreeing to the use of Cookies. You can change your settings at any time. Cookie Policy.


Blog: Dan E. Linstedt Subscribe to this blog's RSS feed!

Dan Linstedt

Bill Inmon has given me this wonderful opportunity to blog on his behalf. I like to cover everything from DW2.0 to integration to data modeling, including ETL/ELT, SOA, Master Data Management, Unstructured Data, DW and BI. Currently I am working on ways to create dynamic data warehouses, push-button architectures, and automated generation of common data models. You can find me at Denver University where I participate on an academic advisory board for Masters Students in I.T. I can't wait to hear from you in the comments of my blog entries. Thank-you, and all the best; Dan Linstedt http://www.COBICC.com, danL@danLinstedt.com

About the author >

Cofounder of Genesee Academy, RapidACE, and BetterDataModel.com, Daniel Linstedt is an internationally known expert in data warehousing, business intelligence, analytics, very large data warehousing (VLDW), OLTP and performance and tuning. He has been the lead technical architect on enterprise-wide data warehouse projects and refinements for many Fortune 500 companies. Linstedt is an instructor of The Data Warehousing Institute and a featured speaker at industry events. He is a Certified DW2.0 Architect. He has worked with companies including: IBM, Informatica, Ipedo, X-Aware, Netezza, Microsoft, Oracle, Silver Creek Systems, and Teradata.  He is trained in SEI / CMMi Level 5, and is the inventor of The Matrix Methodology, and the Data Vault Data modeling architecture. He has built expert training courses, and trained hundreds of industry professionals, and is the voice of Bill Inmons' Blog on http://www.b-eye-network.com/blogs/linstedt/.

We could learn A LOT about information modeling from the nano molecular levels if we only paid attention. Self-assembly at the nanoscale provides many clues about how we should model our information systems as they grow. This blog entry highlights self assembly and its attributes: repeatable, consistent, and reliable.

"Although man's understanding of how to build and control molecular machines is still at an early stage, nanoscale science and engineering could have a life-enhancing impact on human society comparable in extent to that of electricity, the steam engine, the transistor and the Internet." -- Professor David Leigh, Edinburgh University


ComputerWorld reports that self assembly and mixed silicon circuits are 5 to 7 years off. However they do present some very interesting findings from the lead laboratories in the nation. Here we explore impacts, and apply the ideas to a cross-field: data modeling.

"The neat thing about SAMs is they're very well ordered," McGimpsey says. A field of these SAMs protrudes from the substrate at a well-defined angle—like a small patch of thick, well-tended grass—and can perform several duties, such as improving conductivity or increasing surface area. Such order, McGimpsey says, "means predictability of structure, and thus of properties." (from the ComputerWorld article mentioned earlier)

The order means predictable structure and properties - shouldn't we be taking our data modeling queues from nature? Our current data modeling efforts inside RDBMS engines is ancient history, 3rd normal form has only a few ties to natural structure. Our data models must reflect the natural models at the nanoscale. They need to be repeatable, predictable, and redundant. This is a foundation of the Nanohouse. See my web site for more information.

What does Order bring to the table?
Redundancy, fault-tolerance, control, scalability, repeatability are all attributes of order. If we can provide an ordered data model for our information systems (one that resonates with natural models) we can begin predicting how it will act under certain circumstances. We can also begin producing (automatically) the models that will house our data.

No matter how large the data sets grow, we can always predict exactly how it will perform – especially through the use of Fourier transforms and mathematical formulas. "Natural systems form nano-scale structures," Natural systems also provide accurate accounts of form and function. Why then do we in IT insist on creating artificial modeling elements in a 2 dimensional world to house our data? We should be solely focused on 3D modeling capabilities with repeatable and redundant design (ordered systems).

With IT moving toward SOA, we should also be focusing on the data model behind the scenes – can it self-assemble in the future? Can self-assembly mean self-maintaining data models? Can data models proactively change according to newly arriving stimuli? Can we teach our modeling systems (in the information industry) like chemical experiments? When will our data modelers finally learn that it’s about the FORM and FUNCTION, not just the data itself?

For now, focusing on the biological aspects of nano self-assembly can bring tremendous gains to the data modeling world – if for nothing else, housing huge quantities of information in an itsy-bitsy space and an ordered and repeatable fashion.

Do you believe Data Modeling needs an overhaul? Sound off!


Posted September 13, 2005 5:59 AM
Permalink | No Comments |

Leave a comment

    
Search this blog
Categories ›
Archives ›
Recent Entries ›