In my project, we are using Hadoop 2
, Spark
, Scala
. Scala
is the programming language and Spark
is using here for analysing. we are using Hive
and HBase
both. I can access all details like file etc. of HDFS
using Hive
.
But my confusions are -
- When I can able to performed all jobs using
Hive
, Then whyHBase
is required to store the data. Is it not an overhead? - What are the functionality of
HIVE
andHBase
? - If we only used Hive, Then what should be the problem?
Can anyone please let me know.