DocumentCode
3580366
Title
The reliability of Big Data
Author
Xing Wu ; Xinxing Liu ; Shuji Dai
Author_Institution
Sch. of Comput. Eng. & Sci., Shanghai Univ., Shanghai, China
fYear
2014
Firstpage
295
Lastpage
299
Abstract
The era of Big Data is underway. Compared to Small Data, machine learning, computerized databases and other modern technologies makes it possible for Big Data to do handle massive data and reveal information in a way that individual bits of data can´t. Random sampling´s accuracy depends on ensuring randomness when collecting data from samples. Analyzing only a limited number of data points means errors may get amplified, which may reduce the accuracy of the overall results potentially. Meanwhile, Big Data gathers and analyzes massive data to produce some excellent results that we could never know when we are limited to smaller quantities, like address various societal ills, offer potential of new insights into diverse fields. But as the data comes from an environment of uncertainty and rapid change, bigger data may not a better data. Increasing the volume of data may lead to inaccuracy, but in return for relaxing the standards of allowable errors, which produces more valuable information and better results. This article elaborates the reliability of Big Data. Based on our analysis we have constructed a model to analyze the reliability of Big Data.
Keywords
Big Data; data analysis; random processes; reliability; sampling methods; Big Data reliability; massive data analysis; massive data handling; random sampling; Accuracy; Big data; Data models; Entropy; Google; Indexes; Reliability; Accuracy; Big Data; Messy; Reliability; Small Data;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Technology and Artificial Intelligence Conference (ITAIC), 2014 IEEE 7th Joint International
Print_ISBN
978-1-4799-4420-0
Type
conf
DOI
10.1109/ITAIC.2014.7065054
Filename
7065054
Link To Document