InterviewSolution
This section includes InterviewSolutions, each offering curated multiple-choice questions to sharpen your knowledge and support exam preparation. Choose a topic below to get started.
| 1. |
__________ will clear the RMStateStore and is useful if past applications are no longer needed.(a) -format-state(b) -form-state-store(c) -format-state-store(d) none of the mentionedThis question was posed to me by my college director while I was bunking the class.This question is from YARN topic in section Working of MapReduce of Hadoop |
|
Answer» CORRECT OPTION is (C) -format-state-store Explanation: -format-state-store FORMATS the RMSTATESTORE. |
|
| 2. |
Which of the following command is used to dump the log container?(a) logs(b) log(c) dump(d) all of the mentionedThis question was posed to me during an internship interview.This is a very interesting question from YARN in portion Working of MapReduce of Hadoop |
|
Answer» The CORRECT option is (a) logs |
|
| 3. |
The updated queue configuration should be a valid one i.e. queue-capacity at each level should be equal to _________(a) 50%(b) 75%(c) 100%(d) 0%I had been asked this question during an interview.I'm obligated to ask this question of YARN in portion Working of MapReduce of Hadoop |
|
Answer» CORRECT answer is (c) 100% The best explanation: Queues cannot be deleted, only the ADDITION of new queues is SUPPORTED. |
|
| 4. |
Users can bundle their Yarn code in a _________ file and execute it using jar command.(a) java(b) jar(c) C code(d) xmlThis question was addressed to me in a job interview.This question is from YARN topic in section Working of MapReduce of Hadoop |
|
Answer» RIGHT option is (b) JAR The BEST EXPLANATION: Usage: yarn jar |
|
| 5. |
The CapacityScheduler has a predefined queue called _________(a) domain(b) root(c) rear(d) all of the mentionedThe question was asked in my homework.My question is from YARN in chapter Working of MapReduce of Hadoop |
|
Answer» The correct answer is (B) ROOT |
|
| 6. |
Point out the wrong statement.(a) The multiple of the queue capacity which can be configured to allow a single user to acquire more resources(b) Changing queue properties and adding new queues is very simple(c) Queues cannot be deleted, only addition of new queues is supported(d) None of the mentionedThis question was posed to me by my school teacher while I was bunking the class.I'm obligated to ask this question of YARN topic in section Working of MapReduce of Hadoop |
|
Answer» Right CHOICE is (d) NONE of the mentioned |
|
| 7. |
The queue definitions and properties such as ________ ACLs can be changed, at runtime.(a) tolerant(b) capacity(c) speed(d) all of the mentionedThe question was posed to me during an interview.I want to ask this question from YARN in division Working of MapReduce of Hadoop |
|
Answer» Correct option is (b) capacity |
|
| 8. |
Point out the correct statement.(a) Each queue has strict ACLs which controls which users can submit applications to individual queues(b) Hierarchy of queues is supported to ensure resources are shared among the sub-queues of an organization(c) Queues are allocated a fraction of the capacity of the grid in the sense that a certain capacity of resources will be at their disposal(d) All of the mentionedI had been asked this question during an online exam.The origin of the question is YARN topic in division Working of MapReduce of Hadoop |
|
Answer» RIGHT option is (d) All of the mentioned Explanation: All APPLICATIONS submitted to a QUEUE will have access to the CAPACITY ALLOCATED to the queue. |
|
| 9. |
Yarn commands are invoked by the ________ script.(a) hive(b) bin(c) hadoop(d) homeThe question was asked in quiz.My doubt is from YARN topic in chapter Working of MapReduce of Hadoop |
|
Answer» Right CHOICE is (b) bin |
|
| 10. |
The __________ is responsible for allocating resources to the various running applications subject to familiar constraints of capacities, queues etc.(a) Manager(b) Master(c) Scheduler(d) None of the mentionedThis question was addressed to me in an international level competition.The question is from YARN topic in chapter Working of MapReduce of Hadoop |
|
Answer» The correct choice is (c) Scheduler |
|
| 11. |
The ____________ is the ultimate authority that arbitrates resources among all the applications in the system.(a) NodeManager(b) ResourceManager(c) ApplicationMaster(d) All of the mentionedThe question was posed to me during a job interview.My question is taken from YARN topic in portion Working of MapReduce of Hadoop |
|
Answer» CORRECT ANSWER is (B) ResourceManager The explanation: The ResourceManager and per-node SLAVE, the NodeManager (NM), form the data-computation framework. |
|
| 12. |
MapReduce has undergone a complete overhaul in hadoop is _________(a) 0.21(b) 0.23(c) 0.24(d) 0.26This question was addressed to me in an interview for internship.This is a very interesting question from YARN in division Working of MapReduce of Hadoop |
|
Answer» RIGHT answer is (b) 0.23 Explanation: The fundamental idea of MRv2 is to SPLIT up the two MAJOR functionalities of the JobTracker. |
|
| 13. |
Apache Hadoop YARN stands for _________(a) Yet Another Reserve Negotiator(b) Yet Another Resource Network(c) Yet Another Resource Negotiator(d) All of the mentionedThe question was asked in an international level competition.Question is from YARN topic in portion Working of MapReduce of Hadoop |
|
Answer» The CORRECT CHOICE is (c) YET Another RESOURCE Negotiator |
|
| 14. |
The __________ is a framework-specific entity that negotiates resources from the ResourceManager.(a) NodeManager(b) ResourceManager(c) ApplicationMaster(d) All of the mentionedThis question was posed to me in final exam.I'm obligated to ask this question of YARN topic in portion Working of MapReduce of Hadoop |
|
Answer» Right ANSWER is (c) ApplicationMaster |
|
| 15. |
Point out the wrong statement.(a) From the system perspective, the ApplicationMaster runs as a normal container(b) The ResourceManager is the per-machine slave, which is responsible for launching the applications’ containers(c) The NodeManager is the per-machine slave, which is responsible for launching the applications’ containers, monitoring their resource usage(d) None of the mentionedI got this question at a job interview.My doubt is from YARN topic in division Working of MapReduce of Hadoop |
|
Answer» RIGHT answer is (b) The ResourceManager is the per-machine slave, which is responsible for launching the APPLICATIONS’ containers Explanation: ResourceManager has a scheduler, which is responsible for allocating resources to the VARIOUS applications running in the cluster, according to constraints such as queue CAPACITIES and user limits. |
|
| 16. |
YARN’s dynamic allocation of cluster resources improves utilization over more static _______ rules used in early versions of Hadoop.(a) Hive(b) MapReduce(c) Imphala(d) All of the mentionedThis question was addressed to me during an internship interview.I need to ask this question from YARN in section Working of MapReduce of Hadoop |
|
Answer» The correct OPTION is (B) MapReduce |
|
| 17. |
Point out the correct statement.(a) YARN also extends the power of Hadoop to incumbent and new technologies found within the data center(b) YARN is the central point of investment for Hortonworks within the Apache community(c) YARN enhances a Hadoop compute cluster in many ways(d) All of the mentionedI got this question in an international level competition.The doubt is from YARN in division Working of MapReduce of Hadoop |
|
Answer» Correct ANSWER is (d) All of the mentioned |
|
| 18. |
________ is the architectural center of Hadoop that allows multiple data processing engines.(a) YARN(b) Hive(c) Incubator(d) ChuckwaI got this question in a job interview.Asked question is from YARN topic in chapter Working of MapReduce of Hadoop |
|
Answer» Correct answer is (a) YARN |
|
| 19. |
reduceProgress() gets the progress of the job’s reduce-tasks, as a float between _________(a) 0.0-1.0(b) 1.0-2.0(c) 2.0-3.0(d) None of the mentionedI had been asked this question in exam.This intriguing question originated from MapReduce Job topic in division Working of MapReduce of Hadoop |
|
Answer» CORRECT option is (a) 0.0-1.0 Explanation: mapProgress() is used to GET the progress of the JOB’s map-tasks, as a float between 0.0 and 1.0. |
|
| 20. |
_________ gets the diagnostic messages for a given task attempt.(a) getTaskOutputFilter(Configuration conf)(b) getTaskReports(TaskType type)(c) getTrackingURL()(d) all of the mentionedThe question was asked during an interview for a job.My question comes from MapReduce Job in division Working of MapReduce of Hadoop |
|
Answer» The correct choice is (a) getTaskOutputFilter(Configuration CONF) |
|
| 21. |
__________ get events indicating completion (success/failure) of component tasks.(a) getJobName()(b) getJobState()(c) getPriority()(d) getTaskCompletionEvents(int startFrom)This question was addressed to me during an interview for a job.Asked question is from MapReduce Job topic in division Working of MapReduce of Hadoop |
|
Answer» The CORRECT answer is (d) getTaskCompletionEvents(INT startFrom) |
|
| 22. |
Which of the following method is used to get user-specified job name?(a) getJobName()(b) getJobState()(c) getPriority()(d) all of the mentionedThis question was addressed to me in quiz.Asked question is from MapReduce Job in portion Working of MapReduce of Hadoop |
|
Answer» CORRECT answer is (a) getJobName() To EXPLAIN: getPriority() is USED to GET scheduling info of the job. |
|
| 23. |
Which of the following statement can be used to get the storage policy of a file or a directory?(a) hdfs dfsadmin -getStoragePolicy path(b) hdfs dfsadmin -setStoragePolicy path policyName(c) hdfs dfsadmin -listStoragePolicy path policyName(d) all of the mentionedThe question was asked in semester exam.The origin of the question is MapReduce Job in section Working of MapReduce of Hadoop |
|
Answer» RIGHT choice is (a) hdfs dfsadmin -getStoragePolicy path Easy explanation: REFERS to the path REFERRING to either a directory or a FILE. |
|
| 24. |
Point out the wrong statement.(a) getInstance() creates a new Job with particular cluster(b) getInstance(Configuration conf) creates a new Job with no particular Cluster and a given Configuration(c) getInstance(JobStatus status, Configuration conf) creates a new Job with no particular Cluster and given Configuration and JobStatus(d) all of the mentionedI had been asked this question in quiz.This key question is from MapReduce Job topic in portion Working of MapReduce of Hadoop |
|
Answer» The CORRECT OPTION is (a) getInstance() creates a NEW Job with particular cluster |
|
| 25. |
Point out the correct statement.(a) Mover is not similar to Balancer(b) hdfs dfsadmin -setStoragePolicy puts a storage policy to a file or a directory.(c) addCacheArchive add archives to be localized(d) none of the mentionedThis question was posed to me during an online exam.My question is based upon MapReduce Job topic in portion Working of MapReduce of Hadoop |
|
Answer» Correct choice is (C) addCacheArchive ADD archives to be localized |
|
| 26. |
Which of the following is used to list out the storage policies?(a) hdfs storagepolicies(b) hdfs storage(c) hd storagepolicies(d) all of the mentionedThe question was posed to me by my school teacher while I was bunking the class.Question is from MapReduce Job in section Working of MapReduce of Hadoop |
|
Answer» Right OPTION is (a) HDFS storagepolicies |
|
| 27. |
_________ is adata migration tool added for archiving data.(a) Mover(b) Hiver(c) Serde(d) None of the mentionedThis question was addressed to me during an online interview.Query is from MapReduce Job in division Working of MapReduce of Hadoop |
|
Answer» Right option is (a) MOVER |
|
| 28. |
____________ is used for storing one of the replicas in SSD.(a) Hot(b) Lazy_Persist(c) One_SSD(d) All_SSDI had been asked this question in a job interview.This intriguing question originated from MapReduce Job in portion Working of MapReduce of Hadoop |
|
Answer» Correct option is (c) One_SSD |
|
| 29. |
When a block is warm, some of its replicas are stored in DISK and the remaining replicas are stored in _________(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was posed to me in my homework.This question is from MapReduce Job topic in portion Working of MapReduce of Hadoop |
|
Answer» CORRECT ANSWER is (B) ARCHIVE Easiest explanation: WARM storage POLICY is partially hot and partially cold. |
|
| 30. |
Which of the following storage policy is used for both storage and compute?(a) Hot(b) Cold(c) Warm(d) All_SSDThe question was posed to me in a job interview.Asked question is from MapReduce Job topic in division Working of MapReduce of Hadoop |
|
Answer» CORRECT OPTION is (a) Hot Best explanation: When a block is hot, all replicas are STORED in DISK. |
|
| 31. |
Which of the following is only for storage with limited compute?(a) Hot(b) Cold(c) Warm(d) All_SSDThis question was addressed to me in examination.This interesting question is from MapReduce Job topic in section Working of MapReduce of Hadoop |
|
Answer» The CORRECT choice is (B) COLD |
|
| 32. |
Point out the wrong statement.(a) A Storage policy consists of the Policy ID(b) The storage policy can be specified using the “dfsadmin -setStoragePolicy” command(c) dfs.storage.policy.enabled is used for enabling/disabling the storage policy feature(d) None of the mentionedThis question was posed to me by my school principal while I was bunking the class.My doubt is from MapReduce Job topic in portion Working of MapReduce of Hadoop |
|
Answer» Correct choice is (d) NONE of the mentioned |
|
| 33. |
Which of the following has high storage density?(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was addressed to me in examination.The query is from MapReduce Job in division Working of MapReduce of Hadoop |
|
Answer» Right answer is (b) ARCHIVE |
|
| 34. |
___________ is added for supporting writing single replica files in memory.(a) ROM_DISK(b) ARCHIVE(c) RAM_DISK(d) All of the mentionedThis question was posed to me by my school principal while I was bunking the class.This intriguing question originated from MapReduce Job topic in division Working of MapReduce of Hadoop |
|
Answer» The correct choice is (C) RAM_DISK |
|
| 35. |
__________ storage is a solution to decouple growing storage capacity from compute capacity.(a) DataNode(b) Archival(c) Policy(d) None of the mentionedThis question was posed to me by my college professor while I was bunking the class.The question is from MapReduce Job topic in section Working of MapReduce of Hadoop |
|
Answer» Correct choice is (b) Archival |
|
| 36. |
Point out the correct statement.(a) When there is enough space, block replicas are stored according to the storage type list(b) One_SSD is used for storing all replicas in SSD(c) Hot policy is useful only for single replica blocks(d) All of the mentionedThe question was asked in homework.My question is based upon MapReduce Job topic in chapter Working of MapReduce of Hadoop |
|
Answer» Correct option is (a) When there is ENOUGH SPACE, block replicas are stored ACCORDING to the STORAGE TYPE list |
|
| 37. |
The ____________ requires that paths including and leading up to the directories specified in yarn.nodemanager.local-dirs.(a) TaskController(b) LinuxTaskController(c) LinuxController(d) None of the mentionedI have been asked this question by my college professor while I was bunking the class.I need to ask this question from Security topic in division Working of MapReduce of Hadoop |
|
Answer» The CORRECT choice is (B) LinuxTaskController |
|
| 38. |
___________ used by YARN framework which defines how any container launched and controlled.(a) Container(b) ContainerExecutor(c) Executor(d) All of the mentionedThe question was asked in quiz.This intriguing question originated from Security topic in chapter Working of MapReduce of Hadoop |
|
Answer» Right choice is (B) ContainerExecutor |
|
| 39. |
The __________ provides a proxy between the web applications exported by an application and an end user.(a) ProxyServer(b) WebAppProxy(c) WebProxy(d) None of the mentionedThis question was posed to me in homework.I need to ask this question from Security topic in section Working of MapReduce of Hadoop |
|
Answer» Correct option is (b) WebAppProxy |
|
| 40. |
In order to turn on RPC authentication in hadoop, set the value of hadoop.security.authentication property to _________(a) zero(b) kerberos(c) false(d) none of the mentionedI got this question by my school teacher while I was bunking the class.My question is taken from Security in division Working of MapReduce of Hadoop |
|
Answer» The correct OPTION is (b) kerberos |
|
| 41. |
Point out the wrong statement.(a) Data transfer protocol of DataNode does not use the RPC framework of Hadoop(b) Apache Oozie which access the services of Hadoop on behalf of end users need to be able to impersonate end users(c) DataNode must authenticate itself by using privileged ports which are specified by dfs.datanode.address and dfs.datanode.http.address(d) None of the mentionedThe question was asked in an interview for internship.My enquiry is from Security topic in chapter Working of MapReduce of Hadoop |
|
Answer» Right option is (d) None of the mentioned |
|
| 42. |
Data transfer between Web-console and clients are protected by using _________(a) SSL(b) Kerberos(c) SSH(d) None of the mentionedThe question was asked by my school principal while I was bunking the class.I would like to ask this question from Security topic in chapter Working of MapReduce of Hadoop |
|
Answer» CORRECT OPTION is (a) SSL The explanation is: AES offers the greatest cryptographic strength and the BEST PERFORMANCE. |
|
| 43. |
The simplest way to do authentication is using _________ command of Kerberos.(a) auth(b) kinit(c) authorize(d) all of the mentionedI had been asked this question during a job interview.My doubt is from Security topic in chapter Working of MapReduce of Hadoop |
|
Answer» The CORRECT answer is (b) kinit |
|
| 44. |
Point out the correct statement.(a) Hadoop does have the definition of group by itself(b) MapReduce JobHistory server run as same user such as mapred(c) SSO environment is managed using Kerberos with LDAP for Hadoop in secure mode(d) None of the mentionedThe question was posed to me in an internship interview.I'm obligated to ask this question of Security in portion Working of MapReduce of Hadoop |
|
Answer» RIGHT answer is (c) SSO environment is managed using Kerberos with LDAP for Hadoop in secure mode The best I can explain: You can CHANGE a way of mapping by specifying the name of mapping PROVIDER as a value of hadoop.security.group.mapping. |
|
| 45. |
_________ is useful for iterating the properties when all deprecated properties for currently set properties need to be present.(a) addResource(b) setDeprecatedProperties(c) addDefaultResource(d) none of the mentionedThis question was addressed to me in an interview for job.This is a very interesting question from Hadoop Configuration topic in division Working of MapReduce of Hadoop |
|
Answer» CORRECT option is (B) SETDEPRECATEDPROPERTIES The explanation is: setDeprecatedProperties sets all deprecated properties that are not currently set but have a CORRESPONDING new property that is set. |
|
| 46. |
For running hadoop service daemons in Hadoop in secure mode ___________ principals are required.(a) SSL(b) Kerberos(c) SSH(d) None of the mentionedThe question was asked by my school teacher while I was bunking the class.The question is from Security topic in division Working of MapReduce of Hadoop |
|
Answer» Correct OPTION is (b) Kerberos |
|
| 47. |
________ method adds the deprecated key to the global deprecation map.(a) addDeprecits(b) addDeprecation(c) keyDeprecation(d) none of the mentionedThis question was addressed to me during a job interview.Enquiry is from Hadoop Configuration topic in division Working of MapReduce of Hadoop |
|
Answer» The CORRECT answer is (b) addDeprecation |
|
| 48. |
________ checks whether the given key is deprecated.(a) isDeprecated(b) setDeprecated(c) isDeprecatedif(d) all of the mentionedI had been asked this question in exam.This is a very interesting question from Hadoop Configuration topic in chapter Working of MapReduce of Hadoop |
|
Answer» The CORRECT answer is (a) isDeprecated |
|
| 49. |
_________ method clears all keys from the configuration.(a) clear(b) addResource(c) getClass(d) none of the mentionedThis question was posed to me during an interview.Question is taken from Hadoop Configuration topic in section Working of MapReduce of Hadoop |
|
Answer» RIGHT CHOICE is (a) clear For explanation I would say: GETCLASS is used to get the value of the name property as a CLASS. |
|
| 50. |
Point out the wrong statement.(a) addDeprecations adds a set of deprecated keys to the global deprecations(b) configuration parameters cannot be declared final(c) addDeprecations method is lockless(d) none of the mentionedThe question was posed to me during an interview.I'd like to ask this question from Hadoop Configuration in portion Working of MapReduce of Hadoop |
|
Answer» CORRECT answer is (B) CONFIGURATION parameters cannot be declared final To EXPLAIN: Configuration parameters MAY be declared final. |
|