Hadoop DP Notes
Home
Contact Us
Search
What is Hive on Spark?
Hive contains significant support for Apache Spark, wherein Hive execution is configured to Spark:
hive> set spark.home=/location/to/sparkHome;
hive> set hive.execution.engine=spark;
Hive on Spark supports Spark on yarn mode by default.
Newer Post
Older Post
Home
What are the main components of Mapreduce Job ?
Main driver class which provides job configuration parameters. Mapper class which must extend org.apache.hadoop.mapredu...
TO_CHAR
This will be used to extract various date formats. The available date formats as follows. Syntax: to_char ( date , fo...