Q2

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 1

Q5. The composition of Data deals with? a. Structure of data. b. State of data. c. Sensitivity of data.

d. None. Answer :a Q6.The important 3V’s in big data are ? a.Volume, Vulnerability, Variety.
b.Volume, Velocity, Variety. c.Velocity, Vulnerability, Variety. d.Volume, Vulnerability, Velocity.
Answer :b Q7. The data from the CCTV’s coverage and weather forecast report is of data?
a.Structured. b.Unstructured. c.Semi- Structured. d.None. Answer :b Q8. The sensor data, Mach

Big Data vs Data Warehouse

Big Data has become the reality of doing business for organizations today. There is a boom in the
amount of structured as well as raw data that floods every organization daily. If this data is managed
well, it can lead to powerful insights and quality decision making. Big data analytics is the process of
examining large data sets containing a variety of data types to discover some knowledge in
databases, to identify interesting patterns and establish relationships to solve problems, market
trends, customer preferences, and other useful information. Companies and businesses that
implement Big Data Analytics often reap several business benefits. Companies implement Big Data
Analytics because they want to make more informed business decisions. A data warehouse (DW) is a
collection of corporate information and data derived from operational systems and external data
sources. A data warehouse is designed to support business decisions by allowing data consolidation,
analysis and reporting at different aggregate levels. Data is populated into the Data Warehouse
through the processes of extraction, transformation and loading (ETL tools). Data analysis tools, such
as business intelligence software, access the data within the warehouse. Hadoop Environment Big
Data Analytics Hadoop is changing the perception of handling Big Data especially the unstructured
data. Let’s know how Apache Hadoop software library, which is a framework, plays a vital role in
handling Big Data. Apache Hadoop enables surplus data to be streamlined for any distributed
processing system across clusters of computers using simple programming models. It truly is made
to scale up from single servers to a large number of machines, each and every offering local
computation, and storage space. Instead of depending on hardware to provide high-availability, the
library itself is built to detect and handle breakdowns at the application layer, so providing an
extremely available service along with a cluster of computers, as both versions might be vulnerable
to failures. Hadoop Community Package Consists of  File system and OS level abstractions  A
MapReduce engine (either MapReduce or YARN)  The Hadoop Distributed File System (HDFS)  Java
ARchive (JAR) files  Scripts needed to start Hadoop  Source code, documentation and a
contribution section

You might also like