Design your Modern Data Architecture by Making These Six Fundamental Shifts
To succeed in today’s environment, businesses need to lead through growing complexity and volatility, drive operational excellence and enable interaction across enterprise functions, develop higher quality leadership and talent, manage change and unlock new possibilities grounded in data. Organizations are under ever-greater demands to innovate faster and at a larger scale. Data and analytics are being leveraged to generate radical new business models and disrupt traditional industry structures as a result of this response. How are data and analytics driving new ideas and new innovation, and what can executives do to incorporate data-driven innovation more directly into their strategies and initiatives?
Innovation, and the value that it can bring, does not occur in a vacuum. Data is key to motivating and driving innovation.
Analytics, business intelligence (BI), and data management can help organizations innovate, including by making inter-dependencies between humans, institutions, entities, and processes more apparent through study of data relationships. These tools can help organizations obtain a better understanding of how changes to one process or function will affect other processes or functions. However, it will not happen if the organizations don’t have a data architecture to drive innovation and support the business growth. This article is taken from the point of view published by mckinsey and intends to help you to have a clear view about what you can do to drive innovation in your organization.
Data architecture from yesterday can't keep up with today's demands for speed, flexibility, and creativity. Agility is the key to a successful upgrade and a lot of possible benefits. During the last few years, organizations have had to act quickly to deploy new data technologies alongside legacy infrastructure to drive market-driven innovations such as personalized offers, real-time alerts, and predictive maintenance. These technical additions, from data lakes to customer analytics platforms to stream processing, have greatly increased the complexity of data architectures, making it difficult for organizations to deliver new capabilities, maintain existing infrastructures, and ensure the integrity of artificial intelligence (AI) models.
Such slowdowns aren't possible in the current market environment. Leaders like Amazon, Microsoft and Google have been using artificial intelligence to disrupt traditional business models. Cloud providers have introduced cutting-edge products such as serverless data platforms that can be deployed quickly, allowing early adopters to benefit from a shorter time to market and more agility. Users of analytics want more integrated tools, such as automated model deployment platforms, so they can employ new models more quickly.
Many companies are using application programming interfaces (APIs) to expose data from various systems to their data lakes and incorporate insights into front-end apps quickly. Now, companies' need for flexibility and speed has increased as they manage the enormous humanitarian crisis created by the COVID-19 pandemic and prepare themselves for the new normal.
For companies to build a competitive edge or even to maintain parity, they will need a new approach to defining, implementing, and integrating their data stacks, leveraging both cloud and new concepts and components.
This way, there are Six Key Moves to make in order to build a Game-changing Data Architecture. This enables more rapid delivery of new capabilities and vastly simplifies existing architectural approaches.
They involve almost every aspect of data management, including acquisition, processing, storage, analysis, and exposure. Although organizations can implement some adjustment while leaving their core technology stack intact, many require careful re-architecting of the existing data platform and infrastructure, including legacy and newer technologies previously bolted on.
Organizations must have a clear strategic strategy, and data and technology leaders must make courageous decisions to prioritize those changes that will have the most impact on business goals and to invest in the appropriate level of architectural sophistication. As a result, data-architecture blueprints often look very different from one company to another.
The return on investment can be substantial if done correctly. Benefits can be derived from a variety of sources: IT cost savings, productivity improvements, reduced regulatory and operational risk, and the delivery of wholly new capabilities, services, and even entire businesses.
So, what are the major changes and adjustments that enterprises must consider?
1. Cloud-based data platforms (Old-fashioned: On-premise)
Cloud is the most disruptive driver of a new data-architecture strategy, as it allows businesses to rapidly scale AI tools and capabilities for competitive advantage. Major cloud service providers such as AWS, Azure and GCP have revolutionized the way organizations deploy, and run data infrastructure, platforms, and applications at scale.
To modularize application capabilities, for example, an organization can integrate a cloud-based data platform with container technology, which holds microservices such as searching billing data or adding new features to the account. This allows the company to roll out new self-service capabilities to more business customers in days rather than months, deliver large amounts of real-time inventory and transaction data to end users for analytics, and decrease the costs by "buffering" transactions in the cloud rather than on more expensive on-premise systems.
Practical Hints:
2. Real-time data processing (Old-fashioned: Batch)
Real-time data messaging and streaming capabilities have become much less expensive, opening the door for widespread adoption. These technologies enable plenty of new business applications: transportation companies, for example, can provide customers with accurate-to-the-second arrival predictions as their taxi approaches; insurance companies can use real-time behavioral data from smart devices to customize rates; and manufacturers can forecast infrastructure issues based on real-time sensor data.
Real-time streaming functionalities, such as a subscription mechanism, enable data consumers, including data marts and data-driven employees, to subscribe to "topics" in order to receive a constant feed of the transactions they require. In most cases, a shared data lake serves as the "brain" for such services, storing all detailed transactions.
Practical Hints:
3. Modular platforms (Old-fashioned: Pre-integrated commercial solutions)
Businesses must frequently go beyond the limitations of legacy data ecosystems provided by large solution vendors to scale applications. Many organizations are increasingly adopting a highly modular data architecture that incorporates best-of-breed and, in many cases, open-source components that can be replaced with new technologies as needed without affecting other aspects of the data architecture.
This strategy can be used by a company to swiftly deploy new, data-heavy digital services to millions of users and to integrate cloud-based apps at scale. For example, it provides accurate daily views of consumer behavior. The firm can establish an independent data layer that comprises both commercial databases and open-source components. Data is synced with back-end systems via a proprietary enterprise service bus, and business logic on the data is implemented by microservices hosted in containers.
领英推荐
Practical Hints:
4. Decoupled data access (Old-fashioned: Point-to-point)
Data can be exposed via APIs to limit and secure direct access to view and alter data while also providing faster, up-to-date access to common data sets. This enables data to be easily reused across teams, expediting access and facilitating seamless cooperation among analytics teams, allowing for more rapid development of AI use cases in an efficient way.
Instead of depending on proprietary interfaces, one pharmaceutical business is setting up an internal "data marketplace" for all employees using APIs to simplify and standardize access to core data assets.
Practical Hints:
5. Domain-based architecture (Old-fashioned: Enterprise warehouse)
To enhance time to market for new data products and services, several data-architecture leaders have shifted away from a central enterprise data lake toward "domain-driven" solutions that can be customized and "fit for purpose".? While the data sets may still exist on the same physical platform, "product owners" in each business domain (for example, marketing, sales, manufacturing, and so on) are made responsible with organizing their data sets in an easily consumable manner for both internal users and downstream data consumers in other business domains. This approach requires a careful balance to avoid becoming fragmented and inefficient, but in return it can reduce the time spent up front on building new data models into the lake, often from months to just days, and can be a simpler and more effective choice when mirroring a federated business structure or adhering to regulatory limitations on data mobility.
A distributed domain-based architecture can be employed so that sales and operations employees could provide customer, order, and billing data to data scientists for use in AI models or to customers directly via digital channels. Rather than creating a single central data platform, the firm established logical platforms that are managed by product owners inside the sales and operations departments. Product owners are motivated to promote the use of data for analytics, and to drive adoption, they are doing it through digital channels, forums, and hackathons.
Practical Hints:
6. Flexible, extensible data schemas (Old-fashioned: Rigid data models)
To reduce redundancy, software providers' predefined data models and proprietary data models that satisfy specific business-intelligence objectives are frequently designed in highly normalized schemas with rigid database tables and data elements. While this method is still the standard for reporting and regulatory use cases, it also necessitates extensive lengthy development cycles and a deep understanding of the system for incorporating new data elements or data sources, as any modifications can compromise data integrity.
Companies are transitioning to "schema-light" approaches, which use denormalized data models with fewer physical tables to structure data for maximum performance, to gain greater flexibility and a powerful competitive edge when exploring data or supporting advanced analytics. This approach has a number of advantages, including faster data exploration, more flexibility in storing structured and unstructured data, and reduced complexity, as data leaders no longer need to add additional abstraction layers to query relational data, such as multiple "joins" between highly normalized tables.
Practical Hints:
The Next Step: How To Start
Traditional efforts to define and build toward three-to-five-year goal architectural states are both risky and inefficient since data technologies are rapidly developing. Data and technology leaders will benefit most from adopting procedures that allow them to quickly evaluate and deploy new technologies in order to adapt. In this case, four practices are critical: