hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jarek Jarcec Cecho <jar...@apache.org>
Subject Re: Data mismatch when importing data from Oracle to Hive through Sqoop without an error
Date Sat, 16 Mar 2013 22:56:15 GMT
Hi Ajit,
please accept my apologies for late reply. Please upgrade to latest Sqoop version 1.4.3, we've improved the error message so it should be much easier to triangulate the problem.

Jarcec

On Wed, Mar 13, 2013 at 08:44:24AM +0000, Ajit Kumar Shreevastava wrote:
> Thanks Chalcy and Jarek
> 
> Now, I am facing one new problem with Sqoop export utility.
> 
> I have implemented what both of you suggested in chain mail and my problem is solved for importing the data to Hive.
> 
> Now I have truncate the table in oracle and try to export the data from just populated table in Hive is working fine and no record mismatch was found.
> Command for export to oracle is :-->
> [hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop export --connect jdbc:oracle:thin:@10.99.42.11:1521/clouddb --username HDFSUSER  --table BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp_test -P -m 1  --input-fields-terminated-by '\0001' -verbose
> 
> Now I want to create a table in hive as a copy of bttn_bkp_test.
> I described the bttn_bkp_test table and created the table column described in bttn_bkp_test like :-->
> 
> hive> CREATE TABLE IF NOT EXISTS BTTN_BKP(
>     > bttn_id                 double,
>     > data_inst_id            double,
>     > scr_id                  double,
>     > bttn_nu                 double,
>     > cat                     double,
>     > wdth                    double,
>     > hght                    double,
>     > key_scan                double,
>     > key_shft                double,
>     > frgnd_cptn_colr         string,
>     > frgnd_cptn_colr_prsd    string,
>     > bkgd_cptn_colr          string,
>     > bkgd_cptn_colr_prsd     string,
>     > blm_fl                  double,
>     > lclz_fl                 double,
>     > menu_item_nu            double,
>     > bttn_asgn_lvl_id        double,
>     > on_atvt                 double,
>     > on_clik                 double,
>     > enbl_fl                 double,
>     > blm_set_id              double,
>     > bttn_asgn_lvl_name      string,
>     > mkt_id                  double,
>     > crte_ts                 string,
>     > crte_user_id            string,
>     > updt_ts                 string,
>     > updt_user_id            string,
>     > del_ts                  string,
>     > del_user_id             string,
>     > dltd_fl                 double,
>     > menu_item_na            string,
>     > prd_cd                  double,
>     > blm_set_na              string,
>     > sound_file_id           double,
>     > is_dynmc_bttn           double,
>     > frgnd_cptn_colr_id      double,
>     > frgnd_cptn_colr_prsd_id double,
>     > bkgd_cptn_colr_id       double,
>     > bkgd_cptn_colr_prsd_id  double);
> 
> Then I populate this table by following command:-->
> hive> insert OVERWRITE table bttn_bkp
>    > select * from BTTN_BKP_TEST;
> 
> And its insert the same data in bttn_bkp without any error.
> Now I have truncated the oracle table bttn_bkp_test and try to populate the table by bttn_bkp table data of Hive through below Sqoop command :-->
> 
> [hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop export --connect jdbc:oracle:thin:@10.99.42.11:1521/clouddb --username HDFSUSER  --table BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp_test -P -m 1  --input-fields-terminated-by '\0001' -verbose
> 
> And I got below error error:-->
> 
> Warning: /usr/lib/hbase does not exist! HBase imports will fail.
> Please set $HBASE_HOME to the root of your HBase installation.
> 13/03/13 14:00:04 DEBUG tool.BaseSqoopTool: Enabled debug logging.
> Enter password:
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Loaded manager factory: com.cloudera.sqoop.manager.DefaultManagerFactory
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Trying ManagerFactory: com.cloudera.sqoop.manager.DefaultManagerFactory
> 13/03/13 14:00:09 DEBUG manager.DefaultManagerFactory: Trying with scheme: jdbc:oracle:thin:@10.99.42.11
> 13/03/13 14:00:09 DEBUG manager.OracleManager$ConnCache: Instantiated new connection cache.
> 13/03/13 14:00:09 INFO manager.SqlManager: Using default fetchSize of 1000
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Instantiated ConnManager org.apache.sqoop.manager.OracleManager@5557c2bd
> 13/03/13 14:00:09 INFO tool.CodeGenTool: Beginning code generation
> 13/03/13 14:00:09 DEBUG manager.OracleManager: Using column names query: SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0
> 13/03/13 14:00:09 DEBUG manager.OracleManager: Creating a new connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb, using username: HDFSUSER
> 13/03/13 14:00:09 DEBUG manager.OracleManager: No connection paramenters specified. Using regular API for making connection.
> 13/03/13 14:00:10 INFO manager.OracleManager: Time zone has been set to GMT
> 13/03/13 14:00:10 DEBUG manager.SqlManager: Using fetchSize for next query: 1000
> 13/03/13 14:00:10 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0
> 13/03/13 14:00:10 DEBUG manager.OracleManager$ConnCache: Caching released connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: selected columns:
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DATA_INST_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   SCR_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_NU
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CAT
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   WDTH
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   HGHT
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   KEY_SCAN
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   KEY_SHFT
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_FL
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   LCLZ_FL
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MENU_ITEM_NU
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ON_ATVT
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ON_CLIK
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ENBL_FL
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_SET_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MKT_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CRTE_TS
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CRTE_USER_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   UPDT_TS
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   UPDT_USER_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DEL_TS
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DEL_USER_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DLTD_FL
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MENU_ITEM_NA
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   PRD_CD
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_SET_NA
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   SOUND_FILE_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Writing source file: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Table name: BTTN_BKP_TEST
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Columns: BTTN_ID:2, DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2, KEY_SCAN:2, KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12, BKGD_CPTN_COLR:12, BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2, MENU_ITEM_NU:2, BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2, BLM_SET_ID:2, BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93, CRTE_USER_ID:12, UPDT_TS:93, UPDT_USER_ID:12, DEL_TS:93, DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12, PRD_CD:2, BLM_SET_NA:12, SOUND_FILE_ID:2, IS_DYNMC_BTTN:2, FRGND_CPTN_COLR_ID:2, FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2, BKGD_CPTN_COLR_PRSD_ID:2,
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: sourceFilename is BTTN_BKP_TEST.java
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Found existing /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/
> 13/03/13 14:00:10 INFO orm.CompilationManager: HADOOP_HOME is /home/hadoop/hadoop-1.0.3/libexec/..
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Adding source file: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Invoking javac with args:
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -sourcepath
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -d
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -classpath
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> Note: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java uses or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 13/03/13 14:00:11 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.jar
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Scanning for .class files in directory: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Got classfile: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.class -> BTTN_BKP_TEST.class
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Finished writing jar file /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.jar
> 13/03/13 14:00:11 INFO mapreduce.ExportJobBase: Beginning export of BTTN_BKP_TEST
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Using InputFormat: class org.apache.sqoop.mapreduce.ExportInputFormat
> 13/03/13 14:00:11 DEBUG manager.OracleManager$ConnCache: Got cached connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 14:00:11 INFO manager.OracleManager: Time zone has been set to GMT
> 13/03/13 14:00:11 DEBUG manager.OracleManager$ConnCache: Caching released connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/paranamer-2.3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-1.5.3.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/commons-io-1.4.jar
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar
> 13/03/13 14:00:12 INFO input.FileInputFormat: Total input paths to process : 1
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Target numMapTasks=1
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Total input bytes=172704981
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: maxSplitSize=172704981
> 13/03/13 14:00:12 INFO input.FileInputFormat: Total input paths to process : 1
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Generated splits:
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat:   Paths:/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:0+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:67108864+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:134217728+38487253 Locations:NHCLT-PC44-2:;
> 13/03/13 14:00:12 INFO mapred.JobClient: Running job: job_201303121648_0015
> 13/03/13 14:00:13 INFO mapred.JobClient:  map 0% reduce 0%
> 13/03/13 14:00:27 INFO mapred.JobClient: Task Id : attempt_201303121648_0015_m_000000_0, Status : FAILED
> java.lang.NumberFormatException
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>         at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)
> 
> 13/03/13 14:00:33 INFO mapred.JobClient: Task Id : attempt_201303121648_0015_m_000000_1, Status : FAILED
> java.lang.NumberFormatException
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>         at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)
> 
> 13/03/13 14:00:39 INFO mapred.JobClient: Task Id : attempt_201303121648_0015_m_000000_2, Status : FAILED
> java.lang.NumberFormatException
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>         at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)
> 
> 13/03/13 14:00:50 INFO mapred.JobClient: Job complete: job_201303121648_0015
> 13/03/13 14:00:50 INFO mapred.JobClient: Counters: 7
> 13/03/13 14:00:50 INFO mapred.JobClient:   Job Counters
> 13/03/13 14:00:50 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=25370
> 13/03/13 14:00:50 INFO mapred.JobClient:     Total time spent by all reduces waiting after reserving slots (ms)=0
> 13/03/13 14:00:50 INFO mapred.JobClient:     Total time spent by all maps waiting after reserving slots (ms)=0
> 13/03/13 14:00:50 INFO mapred.JobClient:     Launched map tasks=4
> 13/03/13 14:00:50 INFO mapred.JobClient:     Data-local map tasks=4
> 13/03/13 14:00:50 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=0
> 13/03/13 14:00:50 INFO mapred.JobClient:     Failed map tasks=1
> 13/03/13 14:00:50 INFO mapreduce.ExportJobBase: Transferred 0 bytes in 39.0699 seconds (0 bytes/sec)
> 13/03/13 14:00:50 INFO mapreduce.ExportJobBase: Exported 0 records.
> 13/03/13 14:00:50 ERROR tool.ExportTool: Error during export: Export job failed!
> 
> 
> 
> From: Chalcy [mailto:chalcy@gmail.com]
> Sent: Friday, March 08, 2013 11:15 PM
> To: user@sqoop.apache.org
> Cc: user@hive.apache.org
> Subject: Re: Data mismatch when importing data from Oracle to Hive through Sqoop without an error
> 
> Yes, This is newline char in one of the text fields.
> 
> The data is split and forms two rows if the original field has one new line char.  -hive-drop-import-delims is the helper.
> 
> Thanks,
> Chalcy
> On Wed, Mar 6, 2013 at 12:35 PM, Jarek Jarcec Cecho <jarcec@apache.org<mailto:jarcec@apache.org>> wrote:
> Hi Ajit,
> I've seen similar issue many times. Does your table have textual data? If so, can it happen that your textual data contains hive delimiters like new line characters? Because if so then Sqoop might create two lines in for one single row in the table that will be consequently seen as two rows in Hive. As Hive will implicitly convert any invalid values into NULL, it would also explain your NULL values.
> 
> Sqoop offers arguments --hive-drop-import-delims and --hive-delims-replacement to deal with this problem. More information can be found in Sqoop user guide [1].
> 
> Jarcec
> 
> Links:
> 1: http://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_importing_data_into_hive
> 
> On Wed, Mar 06, 2013 at 10:50:26PM +0530, abhijeet gaikwad wrote:
> > If you see Hadoop job counters in your logs - numbers are correct till that
> > point. So normal import in HDFS is working fine. Only reasonable
> > explanation is that there is an issue inserting data in Hive.
> >
> > Have a look at the file on HDFS it should contain the correct number of
> > records with correct data. But if you do "select * ..." from Hive console
> > it will give you NULL values for invalid data, as in the data that cannot
> > be converted to the respective data type of the column of Hive table. This
> > is because Hive imposes data type constraints on the data in files on HDFS.
> > Those extra records is a mystry for me too. I don't think actual file on
> > HDFS has extra records.
> >
> > Sqoop is creating Hive table incorrectly which **may** have caused this
> > issue. Create table manually with correct data types if equivalent
> > available or as string and try loading data. If this works fine we know
> > where the issue is!
> >
> > Thanks,
> > Abhijeet
> > On Wed, Mar 6, 2013 at 2:05 PM, Ajit Kumar Shreevastava <
> > Ajit.Shreevastava@hcl.com<mailto:Ajit.Shreevastava@hcl.com>> wrote:
> >
> >
> > >  Hi Abhijeet,****
> > >
> > > ** **
> > >
> > > Thanks for your response.****
> > >
> > > If values that don't fit in double must be getting inserted as Null is the
> > > case then count should not be mis-match in both the case.****
> > >
> > > Here the null value inserted are extra value apart from the other value
> > > which is already present in both Oracle Table and Hive table.****
> > >
> > > ** **
> > >
> > > Correct me if I am wrong in interpretation.****
> > >
> > > ** **
> > >
> > > Thanks and Regards,****
> > >
> > > Ajit Kumar Shreevastava****
> > >
> > > ** **
> > >
> > > *From:* abhijeet gaikwad [mailto:abygaikwad17@gmail.com<mailto:abygaikwad17@gmail.com>]
> > > *Sent:* Wednesday, March 06, 2013 1:46 PM
> > > *To:* user@hive.apache.org<mailto:user@hive.apache.org>
> > > *Cc:* user@sqoop.apache.org<mailto:user@sqoop.apache.org>
> > > *Subject:* Re: Data mismatch when importing data from Oracle to Hive
> > > through Sqoop without an error****
> > >
> > >   ** **
> > >
> > > Sqoop maps numeric and decimal types (RDBMS) to double (Hive). I think the
> > > values that don't fit in double must be getting inserted as NULL.
> > > You can see this warning in your logs.
> > >
> > > Thanks,
> > > Abhijeet****
> > >
> > > On Wed, Mar 6, 2013 at 1:32 PM, Ajit Kumar Shreevastava <
> > > Ajit.Shreevastava@hcl.com<mailto:Ajit.Shreevastava@hcl.com>> wrote:****
> > >
> > > Hi all,****
> > >
> > > I have notice one interesting thing in the below result-set.****
> > >
> > > I have fired one query in both Oracle and Hive shell and found the
> > > following result set:à****
> > >
> > >  ****
> > >
> > > *SQL> select count(1) from bttn*****
> > >
> > > *  2  where bttn_id is null or data_inst_id is null or scr_id is null;****
> > > *
> > >
> > > * *****
> > >
> > > *  COUNT(1)*****
> > >
> > > *----------*****
> > >
> > > *         0*****
> > >
> > > *hive> select count(1) from bttn*****
> > >
> > > *    > where bttn_id is null or data_inst_id is null or scr_id is null;***
> > > **
> > >
> > > Total MapReduce jobs = 1****
> > >
> > > Launching Job 1 out of 1****
> > >
> > > Number of reduce tasks determined at compile time: 1****
> > >
> > > In order to change the average load for a reducer (in bytes):****
> > >
> > >   set hive.exec.reducers.bytes.per.reducer=<number>****
> > >
> > > In order to limit the maximum number of reducers:****
> > >
> > >   set hive.exec.reducers.max=<number>****
> > >
> > > In order to set a constant number of reducers:****
> > >
> > >   set mapred.reduce.tasks=<number>****
> > >
> > > Starting Job = job_201303051835_0020, Tracking URL =
> > > http://NHCLT-PC44-2:50030/jobdetails.jsp?jobid=job_201303051835_0020<http://nhclt-pc44-2:50030/jobdetails.jsp?jobid=job_201303051835_0020>
> > > ****
> > >
> > > Kill Command = /home/hadoop/hadoop-1.0.3/bin/hadoop job  -kill
> > > job_201303051835_0020****
> > >
> > > Hadoop job information for Stage-1: number of mappers: 1; number of
> > > reducers: 1****
> > >
> > > 2013-03-06 13:22:56,908 Stage-1 map = 0%,  reduce = 0%****
> > >
> > > 2013-03-06 13:23:05,928 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:06,931 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:07,934 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:08,938 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:09,941 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:10,944 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:11,947 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:12,956 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:13,959 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:14,962 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:15,965 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:16,969 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU
> > > 5.2 sec****
> > >
> > > 2013-03-06 13:23:17,974 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:18,977 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:19,981 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:20,985 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:21,988 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:22,995 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > 2013-03-06 13:23:23,998 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 6.95 sec****
> > >
> > > MapReduce Total cumulative CPU time: 6 seconds 950 msec****
> > >
> > > Ended Job = job_201303051835_0020****
> > >
> > > MapReduce Jobs Launched:****
> > >
> > > Job 0: Map: 1  Reduce: 1   Cumulative CPU: 6.95 sec   HDFS Read: 184270926
> > > HDFS Write: 4 SUCCESS****
> > >
> > > Total MapReduce CPU Time Spent: 6 seconds 950 msec****
> > >
> > > OK****
> > >
> > > *986*****
> > >
> > > Time taken: 35.983 seconds****
> > >
> > > *hive>*****
> > >
> > >  ****
> > >
> > > and  *739169 - 738183=986*****
> > >
> > > * *****
> > >
> > > *can anyone tell me why this happened as BTTN_ID  ,  DATA_INST_ID,
> > > SCR_ID   are not null constrains of BTTN table and also composite Primary
> > > Key of the table?*****
> > >
> > > *Also tell me how can I prevent this unnecessary data generation in HIVE
> > > table.*****
> > >
> > > * *****
> > >
> > > *Regards*****
> > >
> > > *Ajit Kumar Shreevastava*****
> > >
> > > * *****
> > >
> > > *From:* Ajit Kumar Shreevastava
> > > *Sent:* Wednesday, March 06, 2013 12:40 PM
> > > *To:* 'user@sqoop.apache.org<mailto:user@sqoop.apache.org>'
> > > *Cc:* user@hive.apache.org<mailto:user@hive.apache.org>
> > > *Subject:* Data mismatch when importing data from Oracle to Hive through
> > > Sqoop without an error
> > > ****
> > >
> > >    ****
> > >
> > > HI,****
> > >
> > >  ****
> > >
> > > I have a table BTTN in Oracle database having *738183* records:à****
> > >
> > >  ****
> > >
> > > *SQL> select count(1) from bttn;*****
> > >
> > >  ****
> > >
> > > *  COUNT(1)*****
> > >
> > > *----------*****
> > >
> > > *    738183*****
> > >
> > > * *****
> > >
> > > Now I want to import this table to HIVE  and I have fired the following
> > > command at command prompt:à****
> > >
> > >  ****
> > >
> > > *[hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop import --connect
> > > jdbc:oracle:thin:@10.99.42.11:1521/clouddb<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb> --username HDFSUSER  --table
> > > BTTN --verbose -P --hive-table bttn --create-hive-table --hive-import
> > > --hive-home /home/hadoop/user/hive/warehouse*****
> > >
> > > *The output is as follows:**à*****
> > >
> > > Warning: /usr/lib/hbase does not exist! HBase imports will fail.****
> > >
> > > Please set $HBASE_HOME to the root of your HBase installation.****
> > >
> > > 13/03/06 12:16:25 DEBUG tool.BaseSqoopTool: Enabled debug logging.****
> > >
> > > Enter password:****
> > >
> > > 13/03/06 12:16:32 INFO tool.BaseSqoopTool: Using Hive-specific delimiters
> > > for output. You can override****
> > >
> > > 13/03/06 12:16:32 INFO tool.BaseSqoopTool: delimiters with
> > > --fields-terminated-by, etc.****
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Loaded manager factory:
> > > com.cloudera.sqoop.manager.DefaultManagerFactory****
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Trying ManagerFactory:
> > > com.cloudera.sqoop.manager.DefaultManagerFactory****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.DefaultManagerFactory: Trying with scheme:
> > > jdbc:oracle:thin:@10.99.42.11<mailto:jdbc%3Aoracle%3Athin%3A@10.99.42.11>****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager$ConnCache: Instantiated new
> > > connection cache.****
> > >
> > > 13/03/06 12:16:32 INFO manager.SqlManager: Using default fetchSize of 1000
> > > ****
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Instantiated ConnManager
> > > org.apache.sqoop.manager.OracleManager@2393385d****<mailto:org.apache.sqoop.manager.OracleManager@2393385d****>
> > >
> > > 13/03/06 12:16:32 INFO tool.CodeGenTool: Beginning code generation****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: Using column names query:
> > > SELECT t.* FROM BTTN t WHERE 1=0****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: Creating a new connection
> > > for jdbc:oracle:thin:@10.99.42.11:1521/clouddb<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb>, using username: HDFSUSER**
> > > **
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: No connection paramenters
> > > specified. Using regular API for making connection.****
> > >
> > > 13/03/06 12:16:32 INFO manager.OracleManager: Time zone has been set to GMT
> > > ****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.SqlManager: Using fetchSize for next
> > > query: 1000****
> > >
> > > 13/03/06 12:16:32 INFO manager.SqlManager: Executing SQL statement: SELECT
> > > t.* FROM BTTN t WHERE 1=0****
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager$ConnCache: Caching released
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: selected columns:****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DATA_INST_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   SCR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_NU****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CAT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   WDTH****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   HGHT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   KEY_SCAN****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   KEY_SHFT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   LCLZ_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MENU_ITEM_NU****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ON_ATVT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ON_CLIK****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ENBL_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_SET_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MKT_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CRTE_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CRTE_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   UPDT_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   UPDT_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DEL_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DEL_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DLTD_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MENU_ITEM_NA****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   PRD_CD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_SET_NA****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   SOUND_FILE_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Writing source file:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Table name: BTTN****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Columns: BTTN_ID:2,
> > > DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2, KEY_SCAN:2,
> > > KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12, BKGD_CPTN_COLR:12,
> > > BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2, MENU_ITEM_NU:2,
> > > BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2, BLM_SET_ID:2,
> > > BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93, CRTE_USER_ID:12, UPDT_TS:93,
> > > UPDT_USER_ID:12, DEL_TS:93, DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12,
> > > PRD_CD:2, BLM_SET_NA:12, SOUND_FILE_ID:2, IS_DYNMC_BTTN:2,
> > > FRGND_CPTN_COLR_ID:2, FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2,
> > > BKGD_CPTN_COLR_PRSD_ID:2,****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: sourceFilename is BTTN.java****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Found existing
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/****
> > >
> > > 13/03/06 12:16:32 INFO orm.CompilationManager: HADOOP_HOME is
> > > /home/hadoop/hadoop-1.0.3/libexec/..****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Adding source file:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Invoking javac with args:*
> > > ***
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -sourcepath****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -d****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -classpath****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:
> > > /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> > > ****
> > >
> > > Note: /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java
> > > uses or overrides a deprecated API.****
> > >
> > > Note: Recompile with -Xlint:deprecation for details.****
> > >
> > > 13/03/06 12:16:34 INFO orm.CompilationManager: Writing jar file:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Scanning for .class files
> > > in directory: /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e**
> > > **
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Got classfile:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.class ->
> > > BTTN.class****
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Finished writing jar file
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set to GMT
> > > ****
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching released
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 WARN manager.OracleManager: The table BTTN contains a
> > > multi-column primary key. Sqoop will default to the column BTTN_ID only for
> > > this job.****
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set to GMT
> > > ****
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching released
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 WARN manager.OracleManager: The table BTTN contains a
> > > multi-column primary key. Sqoop will default to the column BTTN_ID only for
> > > this job.****
> > >
> > > 13/03/06 12:16:34 INFO mapreduce.ImportJobBase: Beginning import of BTTN**
> > > **
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set to GMT
> > > ****
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching released
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.DataDrivenImportJob: Using table class:
> > > BTTN****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.DataDrivenImportJob: Using InputFormat:
> > > class com.cloudera.sqoop.mapreduce.db.OracleDataDrivenDBInputFormat****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/ojdbc6.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/ojdbc6.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/paranamer-2.3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/avro-1.5.3.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/commons-io-1.4.jar****
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:
> > > file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar****
> > >
> > > 13/03/06 12:16:35 INFO db.DataDrivenDBInputFormat: BoundingValsQuery:
> > > SELECT MIN(BTTN_ID), MAX(BTTN_ID) FROM BTTN****
> > >
> > > 13/03/06 12:16:35 INFO mapred.JobClient: Running job: job_201303051835_0014
> > > ****
> > >
> > > 13/03/06 12:16:36 INFO mapred.JobClient:  map 0% reduce 0%****
> > >
> > > 13/03/06 12:16:52 INFO mapred.JobClient:  map 25% reduce 0%****
> > >
> > > 13/03/06 12:17:01 INFO mapred.JobClient:  map 75% reduce 0%****
> > >
> > > 13/03/06 12:17:13 INFO mapred.JobClient:  map 100% reduce 0%****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient: Job complete:
> > > job_201303051835_0014****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient: Counters: 18****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   Job Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=71591****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total time spent by all
> > > reduces waiting after reserving slots (ms)=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total time spent by all maps
> > > waiting after reserving slots (ms)=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Launched map tasks=4****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   File Output Format Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Bytes Written=184266237****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   FileSystemCounters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     HDFS_BYTES_READ=454****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=117870****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=184266237*
> > > ***
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   File Input Format Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Bytes Read=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   Map-Reduce Framework****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Map input records=738183****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Physical memory (bytes)
> > > snapshot=735293440****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Spilled Records=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     CPU time spent (ms)=28950****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total committed heap usage
> > > (bytes)=581435392****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Virtual memory (bytes)
> > > snapshot=6086893568<tel:6086893568>****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Map output records=738183****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     SPLIT_RAW_BYTES=454****
> > >
> > > 13/03/06 12:17:19 INFO mapreduce.ImportJobBase: Transferred 175.73 MB in
> > > 45.4959 seconds (3.8625 MB/sec)****
> > >
> > > 13/03/06 12:17:19 INFO mapreduce.ImportJobBase: Retrieved 738183 records.*
> > > ***
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Hive.inputTable: BTTN****
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Hive.outputTable: bttn****
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager: Using column names query:
> > > SELECT t.* FROM BTTN t WHERE 1=0****
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager$ConnCache: Got cached
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:17:19 INFO manager.OracleManager: Time zone has been set to GMT
> > > ****
> > >
> > > 13/03/06 12:17:19 DEBUG manager.SqlManager: Using fetchSize for next
> > > query: 1000****
> > >
> > > 13/03/06 12:17:19 INFO manager.SqlManager: Executing SQL statement: SELECT
> > > t.* FROM BTTN t WHERE 1=0****
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager$ConnCache: Caching released
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****<http://jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_ID had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DATA_INST_ID had to be
> > > cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column SCR_ID had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_NU had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column CAT had to be cast to a
> > > less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column WDTH had to be cast to
> > > a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column HGHT had to be cast to
> > > a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column KEY_SCAN had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column KEY_SHFT had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BLM_FL had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column LCLZ_FL had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column MENU_ITEM_NU had to be
> > > cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_ASGN_LVL_ID had to
> > > be cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ON_ATVT had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ON_CLIK had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ENBL_FL had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BLM_SET_ID had to be
> > > cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column MKT_ID had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column CRTE_TS had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column UPDT_TS had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DEL_TS had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DLTD_FL had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column PRD_CD had to be cast
> > > to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column SOUND_FILE_ID had to be
> > > cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column IS_DYNMC_BTTN had to be
> > > cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column FRGND_CPTN_COLR_ID had
> > > to be cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column FRGND_CPTN_COLR_PRSD_ID
> > > had to be cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BKGD_CPTN_COLR_ID had
> > > to be cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BKGD_CPTN_COLR_PRSD_ID
> > > had to be cast to a less precise type in Hive****
> > >
> > > 13/03/06 12:17:19 DEBUG hive.TableDefWriter: Create statement: CREATE
> > > TABLE `bttn` ( `BTTN_ID` DOUBLE, `DATA_INST_ID` DOUBLE, `SCR_ID` DOUBLE,
> > > `BTTN_NU` DOUBLE, `CAT` DOUBLE, `WDTH` DOUBLE, `HGHT` DOUBLE, `KEY_SCAN`
> > > DOUBLE, `KEY_SHFT` DOUBLE, `FRGND_CPTN_COLR` STRING, `FRGND_CPTN_COLR_PRSD`
> > > STRING, `BKGD_CPTN_COLR` STRING, `BKGD_CPTN_COLR_PRSD` STRING, `BLM_FL`
> > > DOUBLE, `LCLZ_FL` DOUBLE, `MENU_ITEM_NU` DOUBLE, `BTTN_ASGN_LVL_ID` DOUBLE,
> > > `ON_ATVT` DOUBLE, `ON_CLIK` DOUBLE, `ENBL_FL` DOUBLE, `BLM_SET_ID` DOUBLE,
> > > `BTTN_ASGN_LVL_NAME` STRING, `MKT_ID` DOUBLE, `CRTE_TS` STRING,
> > > `CRTE_USER_ID` STRING, `UPDT_TS` STRING, `UPDT_USER_ID` STRING, `DEL_TS`
> > > STRING, `DEL_USER_ID` STRING, `DLTD_FL` DOUBLE, `MENU_ITEM_NA` STRING,
> > > `PRD_CD` DOUBLE, `BLM_SET_NA` STRING, `SOUND_FILE_ID` DOUBLE,
> > > `IS_DYNMC_BTTN` DOUBLE, `FRGND_CPTN_COLR_ID` DOUBLE,
> > > `FRGND_CPTN_COLR_PRSD_ID` DOUBLE, `BKGD_CPTN_COLR_ID` DOUBLE,
> > > `BKGD_CPTN_COLR_PRSD_ID` DOUBLE) COMMENT 'Imported by sqoop on 2013/03/06
> > > 12:17:19' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES TERMINATED
> > > BY '\012' STORED AS TEXTFILE****
> > >
> > > 13/03/06 12:17:19 DEBUG hive.TableDefWriter: Load statement: LOAD DATA
> > > INPATH 'hdfs://NHCLT-PC44-2:8020/user/hadoop/BTTN' INTO TABLE `bttn`****
> > >
> > > 13/03/06 12:17:19 INFO hive.HiveImport: Removing temporary files from
> > > import process: hdfs://NHCLT-PC44-2:8020/user/hadoop/BTTN/_logs****
> > >
> > > 13/03/06 12:17:19 INFO hive.HiveImport: Loading uploaded data into Hive***
> > > *
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Using external Hive process.****
> > >
> > > 13/03/06 12:17:21 INFO hive.HiveImport: Logging initialized using
> > > configuration in file:/home/hadoop/Hive/conf/hive-log4j.properties****
> > >
> > > 13/03/06 12:17:21 INFO hive.HiveImport: Hive history
> > > file=/home/hadoop/tmp/hadoop/hive_job_log_hadoop_201303061217_790206596.txt
> > > ****
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: OK****
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: Time taken: 11.345 seconds****
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: Loading data to table default.bttn
> > > ****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Table default.bttn stats:
> > > [num_partitions: 0, num_files: 5, num_rows: 0, total_size: 184266237,
> > > raw_data_size: 0]****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: OK****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Time taken: 0.36 seconds****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Hive import complete.****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Export directory is empty,
> > > removing it.****
> > >
> > > *[hadoop@NHCLT-PC44-2 sqoop-oper]$*****
> > >
> > > * *****
> > >
> > > Now I have fired the following query at hive prompt:à****
> > >
> > >  ****
> > >
> > > *hive>* *select count(1) from bttn;*****
> > >
> > > Total MapReduce jobs = 1****
> > >
> > > Launching Job 1 out of 1****
> > >
> > > Number of reduce tasks determined at compile time: 1****
> > >
> > > In order to change the average load for a reducer (in bytes):****
> > >
> > >   set hive.exec.reducers.bytes.per.reducer=<number>****
> > >
> > > In order to limit the maximum number of reducers:****
> > >
> > >   set hive.exec.reducers.max=<number>****
> > >
> > > In order to set a constant number of reducers:****
> > >
> > >   set mapred.reduce.tasks=<number>****
> > >
> > > Starting Job = job_201303051835_0016, Tracking URL =
> > > http://NHCLT-PC44-2:50030/jobdetails.jsp?jobid=job_201303051835_0016<http://nhclt-pc44-2:50030/jobdetails.jsp?jobid=job_201303051835_0016>
> > > ****
> > >
> > > Kill Command = /home/hadoop/hadoop-1.0.3/bin/hadoop job  -kill
> > > job_201303051835_0016****
> > >
> > > Hadoop job information for Stage-1: number of mappers: 1; number of
> > > reducers: 1****
> > >
> > > 2013-03-06 12:21:45,350 Stage-1 map = 0%,  reduce = 0%****
> > >
> > > 2013-03-06 12:21:54,370 Stage-1 map = 76%,  reduce = 0%****
> > >
> > > 2013-03-06 12:21:57,378 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:21:58,381 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:21:59,385 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:00,388 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:01,391 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:02,394 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:03,397 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:04,401 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:05,404 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:06,407 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:07,410 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:08,415 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU
> > > 3.68 sec****
> > >
> > > 2013-03-06 12:22:09,418 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:10,421 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:11,425 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:12,428 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:13,432 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:14,435 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > 2013-03-06 12:22:15,439 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU
> > > 7.37 sec****
> > >
> > > MapReduce Total cumulative CPU time: 7 seconds 370 msec****
> > >
> > > Ended Job = job_201303051835_0016****
> > >
> > > MapReduce Jobs Launched:****
> > >
> > > Job 0: Map: 1  Reduce: 1   Cumulative CPU: 7.37 sec   HDFS Read: 184270926
> > > HDFS Write: 7 SUCCESS****
> > >
> > > Total MapReduce CPU Time Spent: 7 seconds 370 msec****
> > >
> > > OK****
> > >
> > > *739169*****
> > >
> > > Time taken: 37.639 seconds****
> > >
> > > *hive>*****
> > >
> > >  ****
> > >
> > > The data present in *oracle database* is *738183 *but imported into* HIVE
> > > *is* 739169.*****
> > >
> > > Can you please let me know the issue.****
> > >
> > > I have already pasted all my screen log hare.****
> > >
> > > * *****
> > >
> > > *Thanks and Regards,*****
> > >
> > > *Ajit Kumar Shreevastava*****
> > >
> > >
> > > ::DISCLAIMER::
> > >
> > >
> > > ----------------------------------------------------------------------------------------------------------------------------------------------------
> > > ****
> > >
> > > The contents of this e-mail and any attachment(s) are confidential and
> > > intended for the named recipient(s) only.
> > > E-mail transmission is not guaranteed to be secure or error-free as
> > > information could be intercepted, corrupted,
> > > lost, destroyed, arrive late or incomplete, or may contain viruses in
> > > transmission. The e mail and its contents
> > > (with or without referred errors) shall therefore not attach any liability
> > > on the originator or HCL or its affiliates.
> > > Views or opinions, if any, presented in this email are solely those of the
> > > author and may not necessarily reflect the
> > > views or opinions of HCL or its affiliates. Any form of reproduction,
> > > dissemination, copying, disclosure, modification,
> > > distribution and / or publication of this message without the prior
> > > written consent of authorized representative of
> > > HCL is strictly prohibited. If you have received this email in error
> > > please delete it and notify the sender immediately.
> > > Before opening any email and/or attachments, please check them for viruses
> > > and other defects.
> > > ****
> > >
> > >
> > > ----------------------------------------------------------------------------------------------------------------------------------------------------
> > > ****
> > >
> > >  ** **
> 

Mime
View raw message