hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Peyman Mohajerian <mohaj...@gmail.com>
Subject Re: Add few record(s) to a Hive table or a HDFS file on a daily basis
Date Mon, 10 Feb 2014 03:47:35 GMT
The staging table is typically defined as external hive table, data is
loaded directly on HDFS and staging table therefore is able to read that
data directly from HDFS and the transfer it to Hive managed tables, your
current statement. Of course there are variations to this as well.


On Sun, Feb 9, 2014 at 6:45 PM, Raj Hadoop <hadoopraj@yahoo.com> wrote:

>
>
> Hi,
>
> My requirement is a typical Datawarehouse and ETL requirement. I need to
> accomplish
>
> 1) Daily Insert transaction records to a Hive table or a HDFS file. This
> table or file is not a big table ( approximately 10 records per day). I
> don't want to Partition the table / file.
>
>
> I am reading a few articles on this. It was being mentioned that we need
> to load to a staging table in Hive. And then insert like the below :
>
> insert overwrite table finaltable select * from staging;
>
> I am not getting this logic. How should I populate the staging table daily.
>
> Thanks,
> Raj
>
>
>

Mime
View raw message