READ Free Dumps For Cloudera- CCD-410
Question ID 12527 | All keys used for intermediate output from mappers must:
|
Option A | Implement a splittable compression algorithm.
|
Option B | Be a subclass of FileInputFormat.
|
Option C | Implement WritableComparable.
|
Option D | Override isSplitable.
|
Option E | Implement a comparator for speedy sorting.
|
Correct Answer | C |
Explanation Explanation: The MapReduce framework operates exclusively on
pairs, that is, the framework views the input to the job as a set of pairs and produces a set of pairs as the output of the job, conceivably of different types. The key and value classes have to be serializable by the framework and hence need to implement the Writable interface. Additionally, the key classes have to implement the WritableComparable interface to facilitate sorting by the framework. Reference: MapReduce Tutorial
Question ID 12528 | Given a directory of files with the following structure: line number, tab character, string:
Example:
1abialkjfjkaoasdfjksdlkjhqweroij
2kadfjhuwqounahagtnbvaswslmnbfgy
3kjfteiomndscxeqalkzhtopedkfsikj
You want to send each line as one record to your Mapper. Which InputFormat should you
use to complete the line: conf.setInputFormat (____.class) ; ?
|
Option A | SequenceFileAsTextInputFormat
|
Option B | SequenceFileInputFormat
|
Option C | KeyValueFileInputFormat
|
Option D | BDBInputFormat
|
Correct Answer | C |
Explanation Explanation: http://stackoverflow.com/questions/9721754/how-to-parse-customwritable-from-text-in- hadoop