InterviewSolution
| 1. |
Brief about the Job or Application ID. how job history server is handling the Job details and brief about logging and log files. |
|
Answer» After jobs submissions, Job IDs are generated by job tracker in Hadoop 1 and in Hadoop 2/3 Application IDs are generated. Application ID or Job ID is represented as a globally unique identifier for an Application or Job. Example: job_1410450250506_002 / application_1410450250506_002 Task IDs are formed by replacing the job or Application with task prefix within the job Here in the above example, _000002 is the third map task of the job "job_1410450250506_002" Example: attempt_1410450250506_0002_m_0000002_0 When you will open the Job history WEB UI, you will get the image below. Here in the image, you can ABLE to SEE the Job state where the Job is succeeded or Failed. How many Mappers and Reducers are launched whether all the Mappers and Reducers are completed or not you can find all these details. JOB HISTORY Server: When you click the Job id from the Job history server, you will get below image and more or less similar information you will get as above. Overview: Hadoop Counters: This is the most useful option to examine job performance. Hadoop provides several built-in counters as well as you can customize counters as per your requirements. Counters help you to get the below kind of information.
Hadoop counters provide three types of Built-in counters such as :
In addition to this Hadoop provides another 3 counters from other groups by DEFAULT, such as:
File system counters: Under File system counter You can get the information regarding reading and write operations in both the local file system and HDFS as well. The total number of bytes read and written depending upon COMPRESSION algorithms. Here are the few key counters. File_Bytes_Read: The total number of bytes read from the local file system by the map-reduce Tasks. File_Bytes_Write: Total number of bytes written to the local file system. During the Map phase, the mapper task WRITES the intermediate results to the local file system and during the shuffle phase of the Reducer task also write to the local file system when they spill intermediate results to the local file system during sorting.
JOB Counters: You will get Job information related to Mapper and reducer under JOB Counters. The following are the key job counters.
MapReduce Framework counters: You will get all the statistic of MapReduce job under MapReduce framework counter. It will help you to do the performance tuning of the job.
Other counters are as follows:
|
|