Q1
Q1 What is Hadoop primarily used for?
Big data processing
Web hosting
Real-time transaction processing
Network monitoring
Q2
Q2 Which core component of Hadoop is responsible for data storage?
MapReduce
Hive
HDFS
YARN
Q3
Q3 What type of architecture does Hadoop use to process large data sets?
Peer-to-peer
Client-server
Master-slave
Decentralized
Q4
Q4 Hadoop can process data that is:
Structured only
Unstructured only
Semi-structured only
All of the above
Q5
Q5 Which feature of Hadoop makes it suitable for processing large volumes of data?
Fault tolerance
Low cost
Single-threaded processing
Automatic data replication
Q6
Q6 What mechanism does Hadoop use to ensure data is not lost in case of a node failure?
Data mirroring
Data partitioning
Data replication
Data encryption
Q7
Q7 Which programming model is primarily used by Hadoop to process large data sets?
Object-oriented programming
Functional programming
Procedural programming
MapReduce
Q8
Q8 Which command is used to view the contents of a directory in HDFS?
hadoop fs -ls
hadoop fs -dir
hadoop fs -show
hadoop fs -display
Q9
Q9 Which component in Hadoop's architecture is responsible for processing data?
NameNode
DataNode
JobTracker
TaskTracker
Q10
Q10 What role does the NameNode play in Hadoop Architecture?
Manages the cluster's storage resources
Executes user applications
Handles low-level data processing
Serves as the primary data node
Q11
Q11 In Hadoop, what is the function of a DataNode?
Stores data blocks
Processes data blocks
Manages cluster metadata
Coordinates tasks
Q12
Q12 Which type of file system does Hadoop use?
Distributed
Centralized
Virtual
None of the above
Q13
Q13 How does the Hadoop framework handle hardware failures?
Ignoring them
Re-routing tasks
Replicating data
Regenerating data
Q14
Q14 What mechanism allows Hadoop to scale processing capacity?
Adding more nodes to the network
Increasing the storage space on existing nodes
Upgrading CPU speed
Using more efficient algorithms
Q15
Q15 How do you list all nodes in a Hadoop cluster using the command line?
hadoop dfsadmin -report
hadoop fs -ls nodes
hadoop dfs -show nodes
hadoop nodes -list
Q16
Q16 Which command can you use to check the health of the Hadoop file system?
fsck HDFS
hadoop fsck
check HDFS
hdfs check
Q17
Q17 What is the purpose of the hadoop balancer command?
To balance the load on the network
To balance the storage usage across the DataNodes
To upgrade nodes
To restart failed tasks
Q18
Q18 What should you check first if the NameNode is not starting?
Configuration files
DataNode status
HDFS health
Network connectivity
Q19
Q19 When a DataNode is reported as down, what is the first action to take?
Restart the DataNode
Check network connectivity to the DataNode
Delete and reconfigure the DataNode
Perform a full cluster reboot
Q20
Q20 What is a fundamental characteristic of HDFS?
Fault tolerance
Speed optimization
Real-time processing
High transaction rates
Q21
Q21 Which data storage method is used by HDFS to enhance performance and fault tolerance?
Data mirroring
Data replication
Data striping
Data encryption
Q22
Q22 How does HDFS handle very large files?
By breaking them into smaller parts and distributing them
By compressing them
By ignoring them
By storing them on a single node
Q23
Q23 What type of data write operation does HDFS optimize for?
Random writes
Sequential writes
Simultaneous writes
Indexed writes
Q24
Q24 What is the role of the Secondary NameNode in HDFS?
To replace the primary NameNode in case of failure
To take over data node responsibilities
To periodically merge changes to the FS image with the edit log
To store secondary copies of data
Q25
Q25 Which factor influences the block size in HDFS?
The amount of RAM available
The type of data being stored
The total storage capacity of the cluster
The network bandwidth
Q26
Q26 What is the default HDFS command to create a directory?
hadoop fs -mkdir
hadoop fs -createDir
hadoop fs -makeDir
hadoop fs -newDir
Q27
Q27 How do you display the last kilobyte of a file in HDFS?
hadoop fs -tail
hadoop fs -end
hadoop fs -last
hadoop fs -showtail
Q28
Q28 Which command is used to set the replication factor for a file in HDFS?
hadoop fs -setrep
hadoop fs -replicate
hadoop fs -replicationFactor
hadoop fs -setReplication
Q29
Q29 How can you view the list of blocks and their locations for a file in HDFS?
hadoop fsck
hadoop fs -check
hadoop fs -filestatus
hadoop fs -blockinfo
Q30
Q30 What should be the first step if a block of data is missing or corrupt in HDFS?
Run fsck command to identify and fix
Restart the NameNode
Reformat the DataNode
Ignore the error