Friday, 1 May 2015

What is Big Data- Know More about BIG Data

Complexity. Today's data comes from multiple sources. And it is still an undertaking to link, match, cleanse and transform data across systems. However, it is necessary to connect and correlate relationships, hierarchies and multiple data linkages or your data can quickly spiral out of control.

The Importance of Big Data and What You Can Accomplish
The real issue is not that you are acquiring large amounts of data. It's what you do with the data that counts. The hopeful vision is that organizations will be able to take data from any source, harness relevant data and analyze it to find answers that enable 1) cost reductions, 2) time reductions, 3) new product development and optimized offerings, and 4) smarter business decision making. For instance, by combining big data and high-powered analytics, it is possible to:

    Determine root causes of failures, issues and defects in near-real time, potentially saving billions of dollars annually.
    Optimize routes for many thousands of package delivery vehicles while they are on the road.
    Analyze millions of SKUs to determine prices that maximize profit and clear inventory.
    Generate retail coupons at the point of sale based on the customer's current and past purchases.
    Send tailored recommendations to mobile devices while customers are in the right area to take advantage of offers.
    Recalculate entire risk portfolios in minutes.
    Quickly identify customers who matter the most.
    Use clickstream analysis and data mining to detect fraudulent behavior.

 Challenges

 
Many organizations are concerned that the amount of amassed data is becoming so large that it is difficult to find the most valuable pieces of information.

    What if your data volume gets so large and varied you don't know how to deal with it?
    Do you store all your data?
    Do you analyze it all?
    How can you find out which data points are really important?
    How can you use it to your best advantage?

 Until recently, organizations have been limited to using subsets of their data, or they were constrained to simplistic analyses because the sheer volumes of data overwhelmed their processing platforms. But, what is the point of collecting and storing terabytes of data if you can't analyze it in full context, or if you have to wait hours or days to get results? On the other hand, not all business questions are better answered by bigger data. You now have two choices:
A. Incorporate massive data volumes in analysis. If the answers you're seeking will be better provided by analyzing all of your data, go for it. High-performance technologies that extract value from massive amounts of data are here today. One approach is to apply high-performance analytics to analyze the massive amounts of data using technologies such as grid computing, in-database processing and in-memory analytics.

B. Determine upfront which data is relevant. Traditionally, the trend has been to store everything (some call it data hoarding) and only when you query the data do you discover what is relevant. We now have the ability to apply analytics on the front end to determine relevance based on context. This type of analysis determines which data should be included in analytical processes and what can be placed in low-cost storage for later use if needed.

No comments:

Post a Comment