Hello friends in this post we are going to discuss about Big Data Hadoop MCQ | Big Data Hadoop Multiple choice questions | Big Data Hadoop Question Answer | Big Data Hadoop MCQ with answers
If you are looking for more Dumps for MYSKILLZ Visit Here
1) What Is the command tor checking disk usage In hadoop.
- Hadoop fs -disk -space
- Hadoop fs -diskusage
- Hadoop fs -du
- None of the above
Ans: 3
2) How to set the replication factor of a file.
- hadoop fs -setrep -w 3 -R path
- hadoop fs -repset -w 3 -R path
- hadoop fs -setrep -e 3 -R path
- hadoop fs -repset -e 3 -R path
Ans: 1
3) How to set an auto map side Join In hive?
- Set hive.exec.auto.map=true;
- Set hive.auto.convert.Joln=true;
- Set hive.mapred.auto.map.Joln=true;
- Set hive. map. auto, con vert=t rue;
Ans: 2
4) If a database Is having tables with data and you want to delete, then which one Is the correct command.
- Drop database database—name nonrestrict
- Drop database database_name cascade
- Drop schema database-name noncascade
4 Drop database database-name
Ans: 2
5) What Is the default serde used In hive?
- Lazy serdy
- Default serde
- Binary serde
- None of the above.
Ans: 1
6) Create table (Id int, dt string, Ip Int) //1ine1 partitioned by (dt string) //llne2 stored as rcflle; //line 3
- error In line 1;
- error In line 2;
- error In line 3;
- no error;
Ans: 1
7)How can you check your available space and total space in hadoop system?
- HDFS dfsadmin -action
- HDFS dfsadmin -property
- HDFS dfsadmin -report
- None of these
Ans: 3
8) How can you add a catch file to a Job?
- DlstributedCatch.addCatchFlle()
- DistributedCatch.addCatchArchive()
- DistributedCatch.setCatchFiles()
- All of the above.
Ans: 4
9)Which one is not a master daemon?
- Namenode
- Jobtracker
- Tasktracker
- None of these.
Ans: 2
10) Job history is used to support job recovery after a Job tracker restart which parameter you need to set?
- mapred.Jobtracker.restart.recover
- mapred.jobtracker.set recover
- mapred jobtracker restart recover.history
- None of the above
Ans: 1
11) What is TTL In Hbase.
- HBase will automatically delete rows once the expiration time is reached.
- HBase will automatically disable rows once the expiration time is reached.
- It’s just a time taken for executing a job.
- None.
Ans: 1
12) Does HDFS allow appends to files
1 True
2 False
Ans: 1
13) in which file you can Set HBase environment variables.
- hbase-envsh
- hbase-varsh
- hbase-update sh
- None
Ans: 1
14) Which file you need to edit to change rate at which HBase files are rolled and so as the level at which HBase logs messages
- log4j properties
- zookeeper properties
- /ibase properties
- None
Ans: 1
15) What is the default block size in apache HDFS
1 64MB
2 128MB
3 512MB
4 1024MB
Ans: 1
16) What is the default port for jobtracker web Ul?
1 50050
2 50060
3 50070
4 50030
Ans: 4
17) As HDFS works on the principle of
1 Write once Read Many
2 Write Many, Read Many
3 Write Many, Read Once
4 None
Ans: 1
18) Data node decides where to store to data,
1 Yes
2 False
Ans: 2
19) SSH is the communication channel between data node and name node
- Yes 2 False
Ans: 2
20) Reading is parallel and writing is not parallel in HDFS
- True
- False
Ans: 1
21) command to check for various inconsistencies in HDFS
1 FSCK
2 FETCHDT
3 SAFEMODE
4 SAFEANDRECOVE RY
Ans: 1
22) Hive provides
- SQL
- HQL
- PL/SQL
- PL/HQL
Ans: 2
23) HQL Stands for ?
1 Hibernate Query Language
2 Historical Query Language
3. Health Query Language
4 Hive Query Language
Ans: 4
24) HIVE is
- A data mart on hadoop
- A dataware house on hadoop
- a database on hadoop
4 None
Ans: 2
25) HQL allows programmers
1 C# programmers
- Java programmers
- Map-reduce programmers
- python programmers
Ans: 3
26) Hive data is organized into
- Databases
- Tables
3 BucKets/Clusters
4 All of the above
Ans: 4
27) HQL has the statements
1 DDL.DCL
2 DML.TCL
3 DM L, DDL
4 DCL,TCL
Ans: 3
28) The Decimal datatype hasprecision in hive
14
28
3 16
4 N/A
Ans: 4
29) How many bytes takes TINYINT in hive
1.1
2.2
3.4
4.8
Ans: 1
30) regexp_replace(‘sairam’ ‘aijam’) output is
- sal|ram
- sai
- sr
- ram
Ans: 3
31) If explicit conversion falls tnen cast operator returns
- zero
2 one - FALSE
4 Null
Ans: 4
32) which clause can be used to filter rows from a table In HQL
1 group by
2 order by
3 where
4 having
Ans: 3
33) Which one of the following we can use to list the columns and all properties of a -able
1 DECRIBE EXTENDED table_name
2 DECRIBE table_name
3 DECRIBE PROPERTIES tabie_name
4 DECRIBE EXTENDED PROPERTIES taWe__name.
Ans: 1
34) Which clause can be used to restricts the query to a fraction of the buckets in the table rather than the whole table
1 SAMPLE
2 TABLESAMPLE
3 RESTICTTABLE
4 NONE
Ans: 2
35) TABLESAMPLE syntax «
1 TABLESAMPLEfBUCKET X OUT OF(Y»
2 TABLE SAMPLE(BUCKE T X OUT OF y)
3 TABLESAMPLE( BUCKET x IN y)
4 TABLE SAM PL E’BUCKET x «N(y))
Ans: 2
36) How many total default number of dynamic partitions could be created by one DML in hive exec max dynamic partitions parameter
1 10
2 100
3 1000
4 N/A
Ans: 3
37) When using a derby database for a Metastore, how many client instances can connect to Hive?
1 1
2 10
3 Any
4 Cannot Say
Ans: 1
38) in Hadoop 2 0, Name Node High Availability feature is present
1 TRUE
2 FALSE
Ans: 1
39) Name Node is hortizontally scalable due to the facility in Name node Federation
- TRUE
- FALSE
Ans: 1
40) How will you identify when was the last checkpoint done In a cluster
- Using the Name Node Web Ul
- Using the Secondary Name Node Ul
- Using the hadoop dfsadmin -report command
- Using the hadoop fssck command
Ans: 2
41) Hadoop fsck command is used to
1 Check the integrity of the HDFS
2 Check the status of data nodes in the cluster
3 check the status of the NameNode in the cluster
4 Check the status of the Secondary Name Node
Ans: 1
42) How can you determine available HDFS space in your cluster.
1 using hadoop dfsadmin -report command
2 usmg hadoop fsck / command
3 using secondary namenode web Ul
4 Using Data Node Web Ul
Ans: 1
43) A existing Hadoop cluster has 20 slave nodes with quad-core CPUs and 24TB of hard drive space each You plan to add 5 new slave nodes How much disk space can your new nodes contain?
1 Nev/ nodes may have any amount of hard drive space
2 New nodes must have at least 24TB of hard drive space
3 New nodes must have exactly 24TB or hard drive space
4 New nodes must not have more than 24TB of hard drive space
Ans: 2
44) which is a recommended configuration of disk drives for a DataNode
- 10 1TB disk drives in a RAID configuration
- 10 2TB disk drives in a JBOD configuration
- one 3TB disk drive
- 48 2TB disk drives in a RAID configuration
Ans: 2
45) How does the HDFS architecture provide data reliability?
1 Reliance on SAN devices as a DataNode Interface.
- Storing multiple replicas of data blocks on different DataNodes
3 DataNodes make copies of their data blocks, and put them on different local disks
4 Reliance on RAID on each DataNode.
Ans: 2
46) Hcatalog has APIs to connect to HBase
- TRUE
2 FALSE
Ans: 2
47) me path in which the HDFS data will be siored is specified in the lowing me
- hdfs-site xml
- yarn-site xml
3 mapred-site xml
4 core-site xml
Ans: 1
48) To access to a Web user Interface for a specific daemon requires which details
1 The setting for dfs.http address for the NameNode
2 The IP address or DNS/hostname of the NameNode In the cluster
3 The SSL password used to log in to the Hadoop Admin Console
4 The server IP address or DNS/hostname where the daemon is running and the TCP/IP port
Ans: 4
49) What is the default partitioning mechanism?
- Round Robin
- User needs to configure
- Hash Partitioning
- None
Ans: 3
50) is it possible to change the HDFS block size
- TRUE
2 FALSE
Ans: 1
51) Name Node contain
- Meta data, all data blocks
- Metadata and recently used block
- Meta data only
- None of the above
Ans: 3
52) What is variety means to Big Data
- Related data from different source in different formats
- Unrelated data from different source.
Ans: 1
53) Where do you specify the HDFS file system and host location
- hdsf-site.xml
- core-site.xml
- mapred-site.xml
- hive-site.xml
Ans: 2
54) Which file do you use to configure Job Tracker
1 . core-site xml
- mapred-site.xml
- hdfs-site.xml
- Job-tracker xml
Ans: 2
55) Which file is used to define worker nodes
1 . core-site xml
- map red-site, xml
- master-slave xml
4 None
Ans: 4
56) Name Node can toe formatted any time without data loss
1 TRUE
- FALSE
Ans: 2
57) How do you list the files in a HOPS directory
- Is
- Hadoop Is
- hadoop fs -Is
- Hadoop Is -fs
Ans: 3
58) Formatting the Name Node first time will result in
1 . Formats the Name Node disk
- Cleans the HORS data directory
- Just creates the directory structure on the Data Node machine
-4. None of the above
Ans: 4
59) What create the empty directory structure on Name Node
1 . Configure in hdfs-srte xml
- start the Name Node demon
- Format the Name Node
- None of the above
Ans: 3
60) Hadoop answer to Big Data challenge
- Job Tracker and Name Node
- Name Node and Data Node
- Data blocks, keys and value
- HDFS and MapReduce
Ans: 4
61) HDFS Achives High Availability and fault tolerance through
- By spitting files into blocks
- By keeping a copy of frequently accessing data block in Name Node
- By replicating any blocks on multiple data node on the duster
- None of the above
Ans: 3
62) Name node keeps metadata and data files
1 TRUE
- FALSE
Ans: 2
63.Big Data poses challenge to traditional system in terms of
- Network bandwidth!
- Operating system
- Storage and proccessing
- None of the above
Ans: 3
64.What is the function or Secondary* Name Node
1 . Backup to Name Node
2 Helps Name Node in merging fslmage and edit
3. When Name node is busy, it servers the request for the file system
4. None of the above
Ans: 3
65) Hadoop data types are optimized for
-I. Data proccessing
2 Encryption
- Compression
4 Network transmissions
Ans: 4
66) A HDFS rile can be executed
FALSE