Hadoop DP Notes
Home
Contact Us
Search
What is HDFS?
HDFS is a file system designed for storing very large files with streaming data access patterns, running clusters on commodity hardware.
Newer Post
Older Post
Home
What are the main components of Mapreduce Job ?
Main driver class which provides job configuration parameters. Mapper class which must extend org.apache.hadoop.mapredu...
TO_CHAR
This will be used to extract various date formats. The available date formats as follows. Syntax: to_char ( date , fo...