Header menu link for other important links
A Big Data Framework for Quality Assurance and Validation
Published in Blue Eyes Intelligence Engineering and Sciences Engineering and Sciences Publication - BEIESP
Volume: 8
Issue: 2
Pages: 2490 - 2494
Big data is a new technology, which is defined by large amount of data, so it is possible to extract value from the capturing and analysis process. Large data faced many challenges due to various features such as volume, speed, variation, value, complexity and performance. Many organizations face challenges while facing test strategies for structured and unstructured data validation, establishing a proper testing environment, working with non relational databases and maintaining functional testing. These challenges have low quality data in production, delay in execution and increase in cost. Reduce the map for data intensive business and scientific applications Provides parallel and scalable programming model. To get the performance of big data applications, defined as response time, maximum online user data capacity size, and a certain maximum processing capacity. In proposed, to test the health care big data . In health care data contains text file, image file, audio file and video file. To test the big data document, by using two concepts such as big data preprocessing testing and post processing testing. To classify the data from unstructured format to structured format using SVM algorithm. In preprocessing testing test all the data, for the purpose data accuracy. In preprocessing testing such as file size testing, file extension testing and de-duplication testing. In Post Processing to implement the map reduce concept for the use of easily to fetch the data.
About the journal
JournalInternational Journal of Recent Technology and Engineering 2
PublisherBlue Eyes Intelligence Engineering and Sciences Engineering and Sciences Publication - BEIESP
Open Access0