Big Data Volume
Big Data Volume
- Data volume is characterized by the amount of data that is generated continuously.
- Different data types come in different sizes. For example, a blog text is a few kilobytes; voice calls or video files are a few megabytes; sensor data, machine logs, and clickstream data can be in gigabytes.
Example: we can see how the volume aspect of Big Data gets simply overwhelming with the organization Sasstify Inc.
The complexity is not from the type of data but the size too—100 MB per every four hours versus 1 MB per second makes a lot of difference when you look at the amount of compute and associated process cycles.
The most important point to think here is from your organization’s point of view: What are some of the Big Data specifics that can fall into this category and what are the complexities associated with that data?
Let us examine another consumer-oriented corporation and how they looked at this situation within their organization.
- Sasstify, Inc. is a leading photography and videography equipment manufacturer since 1975, providing industry-leading equipment both for commercial and personal use. The company was thriving for over 20 years and was known for its superior customer service.
- Sasstify employed traditional customer relationship management (CRM) techniques to maintain customer loyalty with incentives like club cards, discount coupons, and processing services. With the advent of Web 2.0 and the availability of the Internet, smartphones, and lower-priced competitive offerings, the customer base for Sasstify started declining.
- The traditional decision support platform was able to provide trending, analytics, and KPIs, but was not able to point out any causal analysis. Sasstify lost shares in their customer base and in the stock market.
- The executive management of Sasstify commissioned a leading market research agency to validate the weakness in the data that was used in the decision support platform. The research report pointed out several missing pieces of data that provided insights including sentiment data, data from clickstream analysis, data from online communities, and competitive analysis provided by consumers. Furthermore, the research also pointed to the fact that the company did not have a customer-friendly website and its social media presence was lacking, therefore, its connection with Gen X and Gen Y consumers was near nonexistent.
- Sasstify decided to reinvent the business model from being product-centric to customer-centric. As a part of the makeover, the CRM system was revamped, the customer-facing website was redone, and a strong social media team was formed and tasked with creating connections with Gen X and Gen Y customers. Product research and competitive intelligence were areas of focus with direct reporting to the executive leadership.
- As the business intelligence team started understanding the data requirements for all the new initiatives, it became clear that additional data was needed, and the company had never dealt with this kind of data in its prior life cycle. The additional data sources documented included:
i. Market research reports
ii. Consumer research reports
iii. Survey data
iv. Call center voice calls
v. Emails
vi. Social media data
vii. Excel spreadsheets from multiple business units
viii. Data from interactive web channels
7. The bigger part of the problem was with identifying the content and the context within the new data and aligning it to the enterprise data architecture. In its planning phase, the data warehouse and business intelligence teams estimated the current data to be about 2.5 TB and the new data to be between 2 TB and 3 TB (raw data) per month, which would be between 150 GB and 275 GB post processing. The team decided to adopt to a scalable platform that could handle this volatility with volume of data to be processed, and options included all the Big Data technologies and emerging database technologies.
8. After the implementation cycle, the business intelligence teams across the enterprise were able to use the new platform to successfully plan the business model transformation. The key learning points for the teams included: A new data architecture roadmap and strategy are essential to understand the data, especially considering the volume.
i. Data volume will always be a challenge with Big Data.
ii. Data security will be determined only post processing.
iii. Data acquisition is first and then comes the analysis and discovery.
iv. Data velocity is unpredictable.
v. Non traditional and unorthodox data processing techniques need to be innovated for processing this data type.
vi. Metadata is essential for processing this data successfully.
vii. Metrics and KPIs are key to provide visualization.
viii. Raw data does not need to be stored online for access.
ix. Processed output is needs to be integrated into an enterprise level analytical ecosystem to provide better insights and visibility into the trends and outcomes of business exercises including CRM, Optimization of Inventory, Clickstream analysis and more.
x. The enterprise data warehouse (EDW) is needed for analytics and reporting.
The business model transformation brought with it a tsunami of data that needed to be harnessed and processed for meaningful insights. With a successful change in the data architecture and strategy, Sasstify was able to quickly re-establish itself as a leading provider of photography services including products. With the new business model, the company was able to gain better insights into its legacy and new-generation customer expectations, market trends and their gaps, competition from their view and their customers’ view, and much more. There are nuggets of insights that are found in this extreme volume of information.
The point to pause and ponder is, how might your own organization possibly adapt to new business models? What data might be out there that can help your organization uncover some of these possibilities?.
Please comment your thoughts in comments :)