Explore topic-wise InterviewSolutions in .

This section includes InterviewSolutions, each offering curated multiple-choice questions to sharpen your knowledge and support exam preparation. Choose a topic below to get started.

1.

__________ will clear the RMStateStore and is useful if past applications are no longer needed.(a) -format-state(b) -form-state-store(c) -format-state-store(d) none of the mentionedThis question was posed to me by my college director while I was bunking the class.This question is from YARN topic in section Working of MapReduce of Hadoop

Answer» CORRECT OPTION is (C) -format-state-store

Explanation: -format-state-store FORMATS the RMSTATESTORE.
2.

Which of the following command is used to dump the log container?(a) logs(b) log(c) dump(d) all of the mentionedThis question was posed to me during an internship interview.This is a very interesting question from YARN in portion Working of MapReduce of Hadoop

Answer»

The CORRECT option is (a) logs

Easiest EXPLANATION: USAGE: YARN logs -applicationId .

3.

The updated queue configuration should be a valid one i.e. queue-capacity at each level should be equal to _________(a) 50%(b) 75%(c) 100%(d) 0%I had been asked this question during an interview.I'm obligated to ask this question of YARN in portion Working of MapReduce of Hadoop

Answer» CORRECT answer is (c) 100%

The best explanation: Queues cannot be deleted, only the ADDITION of new queues is SUPPORTED.
4.

Users can bundle their Yarn code in a _________ file and execute it using jar command.(a) java(b) jar(c) C code(d) xmlThis question was addressed to me in a job interview.This question is from YARN topic in section Working of MapReduce of Hadoop

Answer» RIGHT option is (b) JAR

The BEST EXPLANATION: Usage: yarn jar [MAINCLASS] args…
5.

The CapacityScheduler has a predefined queue called _________(a) domain(b) root(c) rear(d) all of the mentionedThe question was asked in my homework.My question is from YARN in chapter Working of MapReduce of Hadoop

Answer»

The correct answer is (B) ROOT

To EXPLAIN I WOULD say: All queues in the system are children of the root queue.

6.

Point out the wrong statement.(a) The multiple of the queue capacity which can be configured to allow a single user to acquire more resources(b) Changing queue properties and adding new queues is very simple(c) Queues cannot be deleted, only addition of new queues is supported(d) None of the mentionedThis question was posed to me by my school teacher while I was bunking the class.I'm obligated to ask this question of YARN topic in section Working of MapReduce of Hadoop

Answer»

Right CHOICE is (d) NONE of the mentioned

The explanation: You need to edit conf/capacity-scheduler.xml and run YARN rmadmin -refreshQueues for changing QUEUE properties.

7.

The queue definitions and properties such as ________ ACLs can be changed, at runtime.(a) tolerant(b) capacity(c) speed(d) all of the mentionedThe question was posed to me during an interview.I want to ask this question from YARN in division Working of MapReduce of Hadoop

Answer»

Correct option is (b) capacity

Best explanation: ADMINISTRATORS can ADD additional QUEUES at runtime, but queues cannot be DELETED at runtime.

8.

Point out the correct statement.(a) Each queue has strict ACLs which controls which users can submit applications to individual queues(b) Hierarchy of queues is supported to ensure resources are shared among the sub-queues of an organization(c) Queues are allocated a fraction of the capacity of the grid in the sense that a certain capacity of resources will be at their disposal(d) All of the mentionedI had been asked this question during an online exam.The origin of the question is YARN topic in division Working of MapReduce of Hadoop

Answer» RIGHT option is (d) All of the mentioned

Explanation: All APPLICATIONS submitted to a QUEUE will have access to the CAPACITY ALLOCATED to the queue.
9.

Yarn commands are invoked by the ________ script.(a) hive(b) bin(c) hadoop(d) homeThe question was asked in quiz.My doubt is from YARN topic in chapter Working of MapReduce of Hadoop

Answer»

Right CHOICE is (b) bin

Explanation: RUNNING the yarn script without any arguments prints the description for all COMMANDS.

10.

The __________ is responsible for allocating resources to the various running applications subject to familiar constraints of capacities, queues etc.(a) Manager(b) Master(c) Scheduler(d) None of the mentionedThis question was addressed to me in an international level competition.The question is from YARN topic in chapter Working of MapReduce of Hadoop

Answer»

The correct choice is (c) Scheduler

The best explanation: The Scheduler is a PURE scheduler in the SENSE that it performs no monitoring or TRACKING of status for the APPLICATION.

11.

The ____________ is the ultimate authority that arbitrates resources among all the applications in the system.(a) NodeManager(b) ResourceManager(c) ApplicationMaster(d) All of the mentionedThe question was posed to me during a job interview.My question is taken from YARN topic in portion Working of MapReduce of Hadoop

Answer» CORRECT ANSWER is (B) ResourceManager

The explanation: The ResourceManager and per-node SLAVE, the NodeManager (NM), form the data-computation framework.
12.

MapReduce has undergone a complete overhaul in hadoop is _________(a) 0.21(b) 0.23(c) 0.24(d) 0.26This question was addressed to me in an interview for internship.This is a very interesting question from YARN in division Working of MapReduce of Hadoop

Answer» RIGHT answer is (b) 0.23

Explanation: The fundamental idea of MRv2 is to SPLIT up the two MAJOR functionalities of the JobTracker.
13.

Apache Hadoop YARN stands for _________(a) Yet Another Reserve Negotiator(b) Yet Another Resource Network(c) Yet Another Resource Negotiator(d) All of the mentionedThe question was asked in an international level competition.Question is from YARN topic in portion Working of MapReduce of Hadoop

Answer»

The CORRECT CHOICE is (c) YET Another RESOURCE Negotiator

The best I can explain: YARN is a cluster management technology.

14.

The __________ is a framework-specific entity that negotiates resources from the ResourceManager.(a) NodeManager(b) ResourceManager(c) ApplicationMaster(d) All of the mentionedThis question was posed to me in final exam.I'm obligated to ask this question of YARN topic in portion Working of MapReduce of Hadoop

Answer»

Right ANSWER is (c) ApplicationMaster

The best I can explain: Each ApplicationMaster has the responsibility for NEGOTIATING APPROPRIATE resource CONTAINERS from the schedule.

15.

Point out the wrong statement.(a) From the system perspective, the ApplicationMaster runs as a normal container(b) The ResourceManager is the per-machine slave, which is responsible for launching the applications’ containers(c) The NodeManager is the per-machine slave, which is responsible for launching the applications’ containers, monitoring their resource usage(d) None of the mentionedI got this question at a job interview.My doubt is from YARN topic in division Working of MapReduce of Hadoop

Answer» RIGHT answer is (b) The ResourceManager is the per-machine slave, which is responsible for launching the APPLICATIONS’ containers

Explanation: ResourceManager has a scheduler, which is responsible for allocating resources to the VARIOUS applications running in the cluster, according to constraints such as queue CAPACITIES and user limits.
16.

YARN’s dynamic allocation of cluster resources improves utilization over more static _______ rules used in early versions of Hadoop.(a) Hive(b) MapReduce(c) Imphala(d) All of the mentionedThis question was addressed to me during an internship interview.I need to ask this question from YARN in section Working of MapReduce of Hadoop

Answer»

The correct OPTION is (B) MapReduce

The best EXPLANATION: Multi-tenant data processing improves an enterprise’s RETURN on its Hadoop investments.

17.

Point out the correct statement.(a) YARN also extends the power of Hadoop to incumbent and new technologies found within the data center(b) YARN is the central point of investment for Hortonworks within the Apache community(c) YARN enhances a Hadoop compute cluster in many ways(d) All of the mentionedI got this question in an international level competition.The doubt is from YARN in division Working of MapReduce of Hadoop

Answer»

Correct ANSWER is (d) All of the mentioned

The explanation: YARN PROVIDES ISVS and developers a consistent FRAMEWORK for writing data access applications that run IN HADOOP.

18.

________ is the architectural center of Hadoop that allows multiple data processing engines.(a) YARN(b) Hive(c) Incubator(d) ChuckwaI got this question in a job interview.Asked question is from YARN topic in chapter Working of MapReduce of Hadoop

Answer»

Correct answer is (a) YARN

Easy EXPLANATION: YARN is the prerequisite for Enterprise HADOOP, providing resource management and a central PLATFORM to deliver consistent operations, SECURITY, and data governance tools across Hadoop clusters.

19.

reduceProgress() gets the progress of the job’s reduce-tasks, as a float between _________(a) 0.0-1.0(b) 1.0-2.0(c) 2.0-3.0(d) None of the mentionedI had been asked this question in exam.This intriguing question originated from MapReduce Job topic in division Working of MapReduce of Hadoop

Answer» CORRECT option is (a) 0.0-1.0

Explanation: mapProgress() is used to GET the progress of the JOB’s map-tasks, as a float between 0.0 and 1.0.
20.

_________ gets the diagnostic messages for a given task attempt.(a) getTaskOutputFilter(Configuration conf)(b) getTaskReports(TaskType type)(c) getTrackingURL()(d) all of the mentionedThe question was asked during an interview for a job.My question comes from MapReduce Job in division Working of MapReduce of Hadoop

Answer»

The correct choice is (a) getTaskOutputFilter(Configuration CONF)

EASIEST explanation: getTaskDiagnostics(TaskAttemptID taskid) gets the diagnostic MESSAGES for a given task ATTEMPT.

21.

__________ get events indicating completion (success/failure) of component tasks.(a) getJobName()(b) getJobState()(c) getPriority()(d) getTaskCompletionEvents(int startFrom)This question was addressed to me during an interview for a job.Asked question is from MapReduce Job topic in division Working of MapReduce of Hadoop

Answer»

The CORRECT answer is (d) getTaskCompletionEvents(INT startFrom)

To EXPLAIN I would say: getPriority() PROVIDES scheduling INFO of the job.

22.

Which of the following method is used to get user-specified job name?(a) getJobName()(b) getJobState()(c) getPriority()(d) all of the mentionedThis question was addressed to me in quiz.Asked question is from MapReduce Job in portion Working of MapReduce of Hadoop

Answer» CORRECT answer is (a) getJobName()

To EXPLAIN: getPriority() is USED to GET scheduling info of the job.
23.

Which of the following statement can be used to get the storage policy of a file or a directory?(a) hdfs dfsadmin -getStoragePolicy path(b) hdfs dfsadmin -setStoragePolicy path policyName(c) hdfs dfsadmin -listStoragePolicy path policyName(d) all of the mentionedThe question was asked in semester exam.The origin of the question is MapReduce Job in section Working of MapReduce of Hadoop

Answer» RIGHT choice is (a) hdfs dfsadmin -getStoragePolicy path

Easy explanation: REFERS to the path REFERRING to either a directory or a FILE.
24.

Point out the wrong statement.(a) getInstance() creates a new Job with particular cluster(b) getInstance(Configuration conf) creates a new Job with no particular Cluster and a given Configuration(c) getInstance(JobStatus status, Configuration conf) creates a new Job with no particular Cluster and given Configuration and JobStatus(d) all of the mentionedI had been asked this question in quiz.This key question is from MapReduce Job topic in portion Working of MapReduce of Hadoop

Answer»

The CORRECT OPTION is (a) getInstance() creates a NEW Job with particular cluster

The EXPLANATION is: getInstance() creates a new Job with particular cluster.

25.

Point out the correct statement.(a) Mover is not similar to Balancer(b) hdfs dfsadmin -setStoragePolicy puts a storage policy to a file or a directory.(c) addCacheArchive add archives to be localized(d) none of the mentionedThis question was posed to me during an online exam.My question is based upon MapReduce Job topic in portion Working of MapReduce of Hadoop

Answer»

Correct choice is (C) addCacheArchive ADD archives to be localized

Easiest explanation: addArchiveToClassPath(Path archive) ADDS an archive path to the current set of classpath entries.

26.

Which of the following is used to list out the storage policies?(a) hdfs storagepolicies(b) hdfs storage(c) hd storagepolicies(d) all of the mentionedThe question was posed to me by my school teacher while I was bunking the class.Question is from MapReduce Job in section Working of MapReduce of Hadoop

Answer»

Right OPTION is (a) HDFS storagepolicies

The explanation: Arguments are NONE for the hdfs storagepolicies COMMAND.

27.

_________ is adata migration tool added for archiving data.(a) Mover(b) Hiver(c) Serde(d) None of the mentionedThis question was addressed to me during an online interview.Query is from MapReduce Job in division Working of MapReduce of Hadoop

Answer»

Right option is (a) MOVER

The best explanation: Mover periodically scans the FILES in HDFS to check if the block PLACEMENT satisfies the STORAGE policy.

28.

____________ is used for storing one of the replicas in SSD.(a) Hot(b) Lazy_Persist(c) One_SSD(d) All_SSDI had been asked this question in a job interview.This intriguing question originated from MapReduce Job in portion Working of MapReduce of Hadoop

Answer»

Correct option is (c) One_SSD

Best EXPLANATION: The remaining replicas are STORED in DISK.

29.

When a block is warm, some of its replicas are stored in DISK and the remaining replicas are stored in _________(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was posed to me in my homework.This question is from MapReduce Job topic in portion Working of MapReduce of Hadoop

Answer» CORRECT ANSWER is (B) ARCHIVE

Easiest explanation: WARM storage POLICY is partially hot and partially cold.
30.

Which of the following storage policy is used for both storage and compute?(a) Hot(b) Cold(c) Warm(d) All_SSDThe question was posed to me in a job interview.Asked question is from MapReduce Job topic in division Working of MapReduce of Hadoop

Answer» CORRECT OPTION is (a) Hot

Best explanation: When a block is hot, all replicas are STORED in DISK.
31.

Which of the following is only for storage with limited compute?(a) Hot(b) Cold(c) Warm(d) All_SSDThis question was addressed to me in examination.This interesting question is from MapReduce Job topic in section Working of MapReduce of Hadoop

Answer»

The CORRECT choice is (B) COLD

Easiest EXPLANATION: When a block is cold, all replicas are stored in the ARCHIVE.

32.

Point out the wrong statement.(a) A Storage policy consists of the Policy ID(b) The storage policy can be specified using the “dfsadmin -setStoragePolicy” command(c) dfs.storage.policy.enabled is used for enabling/disabling the storage policy feature(d) None of the mentionedThis question was posed to me by my school principal while I was bunking the class.My doubt is from MapReduce Job topic in portion Working of MapReduce of Hadoop

Answer»

Correct choice is (d) NONE of the mentioned

Explanation: The EFFECTIVE storage POLICY can be RETRIEVED by the “dfsadmin -getStoragePolicy” command.

33.

Which of the following has high storage density?(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was addressed to me in examination.The query is from MapReduce Job in division Working of MapReduce of Hadoop

Answer»

Right answer is (b) ARCHIVE

To elaborate: Little compute power is ADDED for SUPPORTING ARCHIVAL storage.

34.

___________ is added for supporting writing single replica files in memory.(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was posed to me by my school principal while I was bunking the class.This intriguing question originated from MapReduce Job topic in division Working of MapReduce of Hadoop

Answer»

The correct choice is (C) RAM_DISK

The EXPLANATION is: DISK is the DEFAULT storage type.

35.

__________ storage is a solution to decouple growing storage capacity from compute capacity.(a) DataNode(b) Archival(c) Policy(d) None of the mentionedThis question was posed to me by my college professor while I was bunking the class.The question is from MapReduce Job topic in section Working of MapReduce of Hadoop

Answer»

Correct choice is (b) Archival

The explanation is: Nodes with HIGHER density and LESS expensive storage with LOW compute power are BECOMING available.

36.

Point out the correct statement.(a) When there is enough space, block replicas are stored according to the storage type list(b) One_SSD is used for storing all replicas in SSD(c) Hot policy is useful only for single replica blocks(d) All of the mentionedThe question was asked in homework.My question is based upon MapReduce Job topic in chapter Working of MapReduce of Hadoop

Answer»

Correct option is (a) When there is ENOUGH SPACE, block replicas are stored ACCORDING to the STORAGE TYPE list

The best explanation: The first phase of Heterogeneous Storage changed datanode storage model from a single storage.

37.

The ____________ requires that paths including and leading up to the directories specified in yarn.nodemanager.local-dirs.(a) TaskController(b) LinuxTaskController(c) LinuxController(d) None of the mentionedI have been asked this question by my college professor while I was bunking the class.I need to ask this question from Security topic in division Working of MapReduce of Hadoop

Answer»

The CORRECT choice is (B) LinuxTaskController

Explanation: LinuxTaskController KEEPS track of all paths and DIRECTORIES on datanode.

38.

___________ used by YARN framework which defines how any container launched and controlled.(a) Container(b) ContainerExecutor(c) Executor(d) All of the mentionedThe question was asked in quiz.This intriguing question originated from Security topic in chapter Working of MapReduce of Hadoop

Answer»

Right choice is (B) ContainerExecutor

For explanation: The container process has the same UNIX USER as the NODEMANAGER.

39.

The __________ provides a proxy between the web applications exported by an application and an end user.(a) ProxyServer(b) WebAppProxy(c) WebProxy(d) None of the mentionedThis question was posed to me in homework.I need to ask this question from Security topic in section Working of MapReduce of Hadoop

Answer»

Correct option is (b) WebAppProxy

Explanation: If security is enabled it will warn users before ACCESSING a POTENTIALLY unsafe WEB application. Authentication and authorization using the proxy is handled just like any other PRIVILEGED web application.

40.

In order to turn on RPC authentication in hadoop, set the value of hadoop.security.authentication property to _________(a) zero(b) kerberos(c) false(d) none of the mentionedI got this question by my school teacher while I was bunking the class.My question is taken from Security in division Working of MapReduce of Hadoop

Answer»

The correct OPTION is (b) kerberos

The BEST explanation: SECURITY settings NEED to be modified properly for robustness.

41.

Point out the wrong statement.(a) Data transfer protocol of DataNode does not use the RPC framework of Hadoop(b) Apache Oozie which access the services of Hadoop on behalf of end users need to be able to impersonate end users(c) DataNode must authenticate itself by using privileged ports which are specified by dfs.datanode.address and dfs.datanode.http.address(d) None of the mentionedThe question was asked in an interview for internship.My enquiry is from Security topic in chapter Working of MapReduce of Hadoop

Answer»

Right option is (d) None of the mentioned

Explanation: Authentication is BASED on the assumption that the attacker won’t be ABLE to GET ROOT privileges.

42.

Data transfer between Web-console and clients are protected by using _________(a) SSL(b) Kerberos(c) SSH(d) None of the mentionedThe question was asked by my school principal while I was bunking the class.I would like to ask this question from Security topic in chapter Working of MapReduce of Hadoop

Answer» CORRECT OPTION is (a) SSL

The explanation is: AES offers the greatest cryptographic strength and the BEST PERFORMANCE.
43.

The simplest way to do authentication is using _________ command of Kerberos.(a) auth(b) kinit(c) authorize(d) all of the mentionedI had been asked this question during a job interview.My doubt is from Security topic in chapter Working of MapReduce of Hadoop

Answer»

The CORRECT answer is (b) kinit

To elaborate: HTTP web-consoles should be SERVED by principal DIFFERENT from RPC’s ONE.

44.

Point out the correct statement.(a) Hadoop does have the definition of group by itself(b) MapReduce JobHistory server run as same user such as mapred(c) SSO environment is managed using Kerberos with LDAP for Hadoop in secure mode(d) None of the mentionedThe question was posed to me in an internship interview.I'm obligated to ask this question of Security in portion Working of MapReduce of Hadoop

Answer» RIGHT answer is (c) SSO environment is managed using Kerberos with LDAP for Hadoop in secure mode

The best I can explain: You can CHANGE a way of mapping by specifying the name of mapping PROVIDER as a value of hadoop.security.group.mapping.
45.

_________ is useful for iterating the properties when all deprecated properties for currently set properties need to be present.(a) addResource(b) setDeprecatedProperties(c) addDefaultResource(d) none of the mentionedThis question was addressed to me in an interview for job.This is a very interesting question from Hadoop Configuration topic in division Working of MapReduce of Hadoop

Answer» CORRECT option is (B) SETDEPRECATEDPROPERTIES

The explanation is: setDeprecatedProperties sets all deprecated properties that are not currently set but have a CORRESPONDING new property that is set.
46.

For running hadoop service daemons in Hadoop in secure mode ___________ principals are required.(a) SSL(b) Kerberos(c) SSH(d) None of the mentionedThe question was asked by my school teacher while I was bunking the class.The question is from Security topic in division Working of MapReduce of Hadoop

Answer»

Correct OPTION is (b) Kerberos

Best explanation: Each SERVICE READS authenticate information saved in keytab file with appropriate permission.

47.

________ method adds the deprecated key to the global deprecation map.(a) addDeprecits(b) addDeprecation(c) keyDeprecation(d) none of the mentionedThis question was addressed to me during a job interview.Enquiry is from Hadoop Configuration topic in division Working of MapReduce of Hadoop

Answer»

The CORRECT answer is (b) addDeprecation

For EXPLANATION: addDeprecation does not OVERRIDE any EXISTING entries in the deprecation map.

48.

________ checks whether the given key is deprecated.(a) isDeprecated(b) setDeprecated(c) isDeprecatedif(d) all of the mentionedI had been asked this question in exam.This is a very interesting question from Hadoop Configuration topic in chapter Working of MapReduce of Hadoop

Answer»

The CORRECT answer is (a) isDeprecated

For explanation: Method returns TRUE if the KEY is deprecated and FALSE OTHERWISE.

49.

_________ method clears all keys from the configuration.(a) clear(b) addResource(c) getClass(d) none of the mentionedThis question was posed to me during an interview.Question is taken from Hadoop Configuration topic in section Working of MapReduce of Hadoop

Answer» RIGHT CHOICE is (a) clear

For explanation I would say: GETCLASS is used to get the value of the name property as a CLASS.
50.

Point out the wrong statement.(a) addDeprecations adds a set of deprecated keys to the global deprecations(b) configuration parameters cannot be declared final(c) addDeprecations method is lockless(d) none of the mentionedThe question was posed to me during an interview.I'd like to ask this question from Hadoop Configuration in portion Working of MapReduce of Hadoop

Answer» CORRECT answer is (B) CONFIGURATION parameters cannot be declared final

To EXPLAIN: Configuration parameters MAY be declared final.