Business needs
Facilitates cloud computing and
Hypervisor
Virtualization
D. Virtualization
Business organizations
System development
Employees development
All of the mentioned above
Default location of Hadoop configuration is in $HADOOP /conf/ HOME
If $HADOOP HOME is specified, Sqoop will utilise the default installation location
default location of Hadoop configuration is in $HADOOP HOME/conf/
Sqoop command-line tool serves as a wrapper for the bin/hadoop script that is included with Hadoop as a base.
Spark
HBase
Hive
Pig
Internally managed data
Data feeds from external sources.
It provides access to each and every layer & components of big data stack
All of the mentioned above
Goes beyond the massive volumes
Increasing velocities of data
Both A and B
None of the mentioned above
Data can arrive at fast speed
Enormous datasets can accumulate within very short periods of time
Velocity of data translates into the amount of time it takes for the data to be processed
All of the mentioned above
DBMS
NoSQL
Data store
None of the mentioned above
2
3
4
5
Worth in information
Useless data
Useless information
None of the mentioned above
To visualize the data
To access the data
To process the data
All of the mentioned above
Oozie
Zookepper
MapReduce
All of the mentioned above
TaskReduce
Mapreduce
TaskTracker
JobTracker
HashPartitioner
Map function
Reduce function
All of the mentioned above
Cloud computing
Security management
Integrated approach
None of the mentioned above
Structured format
Unstructured format
Semi-structured format
None of the mentioned above
Hive is a relational database that supports SQL queries.
Pig is a relational database that supports SQL queries.
Both A and B
None of the mentioned above
Exhausting valuable resources on housing data that does not inform business decisions
Spending time sifting through unutilized data sets
Missing out on unique revenue streams and insights
All of the mentioned above
HBase
HDFS
Hive
Mapreduce
Distributed computing
Cluster computing
Parallel computing
All of the mentioned above
Predicted variables
Descriptive variables
Prescriptive variables
All of the mentioned above
Master-slave architecture
Master-worker architecture
Worker-slave architecture
All of the mentioned above
Cloud-based
Data warehouse
System ingestion
All of the mentioned above
TYPE-1 Hypervisor
TYPE- 2 Hypervisor
Both A and B
None of the mentioned above
Indexing and Cataloging, Replication
File Storage and Structure, Query Processing
Transactions Support
All of the mentioned above
Transportation of data from the ingestion layer to the rest of the data pipeline
Data storage
Data identification
None of the mentioned above
Physical machine
Abstract machine
System integration
All of the mentioned above
Webpages
All of the mentioned above
Predictive models
Descriptive models
Decision models
All of the mentioned above
Import data → Clean the data → Develop a predictive model → Integrate the model
Clean the data → Develop a predictive model → Import data → Integrate the model
Clean the data → Develop a predictive model → Import data → Integrate the model
None of the mentioned above
Java
C#
C
C++