Exploring Data Quality: Insights from 'Data Quality Engineering in Financial Services' Book
Recently, I've been immersed in the book "Data Quality Engineering in Financial Services" by Brian Buzzelli making my way through the first couple of chapters. It's been an enlightening journey so far, especially delving into the concept of data as a valuable asset for businesses.
One of the ideas introduced in the book is the concept of Data Quality Specifications (DQS). Essentially, it's about collaborating with process/function experts to define what makes data acceptable both when it's extracted/generated and when it's used by people or other systems.
To achieve this, the book suggests using a framework called "Data Dimensions" to gauge the quality of our data based on certain measures.
In simpler terms, it's about checking our datasets against specific criteria:
By setting these Data Quality Specifications, we're equipped to evaluate how well different datasets measure up.
领英推荐
Here's an example illustrating how we might assess these dimensions:
As I progress through the book, I'm eager to explore how these Data Quality Specifications can refine my own data processing workflows.
Head of Asset Management AI at Manulife | MSCF (MFE) Carnegie Mellon | Quant, Investing, AI, Analytics
11 个月Applications to investment data in particular are important. There are some attributes of investment data that make it materially different from operational data for example.
Business Intelligence and Analytics Architect | Microsoft Azure Data | Microsoft Fabric Engineer + Power BI Engineer | CCH? Tagetik | ERP and CPM Implementation | Microsoft Dynamics 365 ERP Finance and Business Central
11 个月Very good advice from the book. Indeed, companies spending time and effort on data quality is important.