AllExam Dumps

DUMPS, FREE DUMPS, VCP5 DUMPS| VMWARE DUMPS, VCP DUMPS, VCP4 DUMPS, VCAP DUMPS, VCDX DUMPS, CISCO DUMPS, CCNA, CCNA DUMPS, CCNP DUMPS, CCIE DUMPS, ITIL, EXIN DUMPS,


READ Free Dumps For Cloudera- CCD-410





Question ID 12527

All keys used for intermediate output from mappers must:

Option A

Implement a splittable compression algorithm.

Option B

Be a subclass of FileInputFormat.

Option C

 Implement WritableComparable.

Option D

Override isSplitable.

Option E

Implement a comparator for speedy sorting.

Correct Answer C
Explanation Explanation: The MapReduce framework operates exclusively on pairs, that is, the framework views the input to the job as a set of pairs and produces a set of pairs as the output of the job, conceivably of different types. The key and value classes have to be serializable by the framework and hence need to implement the Writable interface. Additionally, the key classes have to implement the WritableComparable interface to facilitate sorting by the framework. Reference: MapReduce Tutorial


Question ID 12528

Given a directory of files with the following structure: line number, tab character, string:
Example:
1abialkjfjkaoasdfjksdlkjhqweroij
2kadfjhuwqounahagtnbvaswslmnbfgy
3kjfteiomndscxeqalkzhtopedkfsikj
You want to send each line as one record to your Mapper. Which InputFormat should you
use to complete the line: conf.setInputFormat (____.class) ; ?

Option A

SequenceFileAsTextInputFormat

Option B

SequenceFileInputFormat

Option C

 KeyValueFileInputFormat

Option D

 BDBInputFormat

Correct Answer C
Explanation Explanation: http://stackoverflow.com/questions/9721754/how-to-parse-customwritable-from-text-in- hadoop