“Is The Data Model Finished Yet?”

“Is The Data Model Finished Yet?”

Naturally this title caught my attention.

Intrigued, I read on.

The article argued that with the advent of Big Data, we no longer need to waste effort putting together often complex and time consuming data models. In this transformed data landscape, everything can be developed rapidly to meet a specific set of requirements. As soon as the next set of requirements arises, if the previous development does not fit them, it will be jettisoned and replaced.

Part of me wondered at this thinking; can the world’s organisations now really operate in this way because the data landscape been so utterly transformed by Big Data?

I remembered back to similar articles I had read over the years. Each repetition of this death knell has occurred with the advent of a major new paradigm on the data landscape. Examples over the years include; the relentless advance of COTS[1] products, the Internet, Agile development, the Cloud and, of course, the Big Data Lake.

But still the data model has survived.

My mind started to sift the evidence that I have personally witnessed. Big Data certainly has had a significant impact on organisations that require data to support their operations, and therefore also on their data models. But I can see no indication that even this seismic shift in data collection and analysis has caused these core organisational definitions to die out.

In fact, for many organisations, a key outcome of adopting Big Data has been the exact opposite; it has resulted in the realisation of their importance. The explosion of technical innovations that have transformed data usage by organisations, has fundamentally altered the way that data models are required to support them in this data-rich environment. Whereas in previous times, they may have been restricted to being viewed as an ‘unwelcome’ but necessary part of development, they are now being recognised for what they truly are; a definition of an organisation's operational lifeblood.

There is now a realisation that data models allow an organisation to ‘know thyself’.

Arguably, they are now more important than ever. It is only with a full and agreed understanding of the ‘What?’, ‘When?’ and ‘How?’ of an organisation’s data structures and flows, that we can contemplate plugging COTS products together, implementing in the Cloud, ingesting data into our Data Lakes, or reporting across the Enterprise system landscape. In the last decade, data models have made the transition from being ad-hoc and limited in scope, to becoming a central pillar of the Enterprise Data Architectural landscape.

Thus, a beneficial by-product from the adoption of Big Data is the realisation that its true benefit cannot be delivered without being able to correlate the meaning from Big Data analysis with the organisation’s Master Data Domains. This has driven the focus for organisations to be able to bring their Master Data under control and thus the processes that manage it.

So, I took what I could from the article, finished my sandwich, and resolved that it was now time to start writing the book that I had always wanted to. Its purpose would be to:

  • Explain the power of data models
  • Describe the easy steps required to define and quality assure them and
  • Define the processes that harness their power to deliver maximum benefit

I finished the book in late 2016. It is entitled "The Data Model Toolkit - Simple Skills To Model The Real World" (ISBN - 13: 978-1782224730) available here in the UK and here in the US.

It is one in a growing popular series including the "Enterprise Data Architecture - How to navigate its landscape" also available from all the usual online stores including here in the UK and here in the US.

 

 

[1] Commercial off the Shelf

Jon Cooke

AI Digital Twins | Simulate business ideas in minutes with AI, real data and Data Object Graphs (DOGs) | Agent DOG Handler | Composable Enterprises with Data Product Pyramid | Data Product Workshop podcast co-host

8 年

The data modelling approach does change in the big data world. The traditional approach is to create enterprise data model would be created up front with elements and attributes being defined across multiple business lines. In the big data world, for structured data, the emphasis is on only joining the data when you need to, ideally on query time e.g schema on read or at least on use-case. This means you spend less time actually modelling ie trying to create large complex Inman or Kimble models up front. You allow the businesses to create their own structures from disparate clusters of data that are not joined but normalised to conform to a common data dictionary.

Liz Rowe, CIPP/US/E/C

SVP, Data Governance, Privacy & Risk

8 年

Agree. It's nott sexy but governance early, governance always

Chris Probert

Partner UK Data Practice Lead at Capco

8 年

Very interesting article. I like the 'knowing your organisation' view. Too many solutions involve consuming large amounts of data and cleaning it up later. This just institutionalised poor behaviours and bad data. Thanks

Dave Knifton

Global Enterprise Data Architect - Data Evangelist | Entrepreneur | Public Speaker | Published Author

8 年

Hi Jeremy - yes could not agree more ...

回复
Jeremy Posner

Experienced Enterprise Data Professional with deep Financial Markets experience | Strategy, Architecture, Governance & Implementation | Modelling and Metadata | Product and Solutions

8 年

Nice piece Dave. Agree entirely. The importance has grown. However, unlike in the past with our old tool-set, the model may not end up being physicalised. Ever. And if it does it's likely to be radically different across data platforms. Which makes a common language for data even MORE important than it was in the old days of hard schemes, as without the model we have a mess. That's a key difference in my eyes.

要查看或添加评论,请登录

Dave Knifton的更多文章

社区洞察

其他会员也浏览了