sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Artem Ervits" <are9...@nyp.org>
Subject RE: SQOOP 1.3.0 questions using SQL Server
Date Fri, 26 Oct 2012 21:48:45 GMT
Don't pay attention to error that this table has two primary keys, I just want to know whether
nvarchar still works.

[xxxxx@SERVERNAME ~]$ clear
[xxxxx@SERVERNAME ~]$ $SQOOP_HOME/bin/sqoop import-all-tables --connect "jdbc:sqlserver://dbserver;username=xxxx;password=xxxxx;database=xxxxx"
--hive-import --create-hive-table -compress  -verbose
12/10/26 17:39:52 DEBUG tool.BaseSqoopTool: Enabled debug logging.
12/10/26 17:39:52 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for output. You
can override
12/10/26 17:39:52 INFO tool.BaseSqoopTool: delimiters with --fields-terminated-by, etc.
12/10/26 17:39:52 ERROR sqoop.ConnFactory: Error loading ManagerFactory information from file
/usr/local/sqoop/conf/managers.d/mssqoop-sqlserver: java.io.IOException: the content of connector
file must be in form of key=value
        at org.apache.sqoop.ConnFactory.addManagersFromFile(ConnFactory.java:219)
        at org.apache.sqoop.ConnFactory.loadManagersFromConfDir(ConnFactory.java:294)
        at org.apache.sqoop.ConnFactory.instantiateFactories(ConnFactory.java:85)
        at org.apache.sqoop.ConnFactory.<init>(ConnFactory.java:62)
        at com.cloudera.sqoop.ConnFactory.<init>(ConnFactory.java:36)
        at org.apache.sqoop.tool.BaseSqoopTool.init(BaseSqoopTool.java:200)
        at org.apache.sqoop.tool.ImportTool.init(ImportTool.java:83)
        at org.apache.sqoop.tool.ImportAllTablesTool.run(ImportAllTablesTool.java:48)
        at org.apache.sqoop.Sqoop.run(Sqoop.java:145)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
        at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)
        at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)
        at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)
        at org.apache.sqoop.Sqoop.main(Sqoop.java:238)
        at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)

12/10/26 17:39:52 DEBUG sqoop.ConnFactory: Loaded manager factory: com.cloudera.sqoop.manager.DefaultManagerFactory
12/10/26 17:39:52 DEBUG sqoop.ConnFactory: Trying ManagerFactory: com.cloudera.sqoop.manager.DefaultManagerFactory
12/10/26 17:39:52 DEBUG manager.DefaultManagerFactory: Trying with scheme: jdbc:sqlserver:
12/10/26 17:39:52 INFO manager.SqlManager: Using default fetchSize of 1000
12/10/26 17:39:52 DEBUG sqoop.ConnFactory: Instantiated ConnManager org.apache.sqoop.manager.SQLServerManager@380e28b9
12/10/26 17:39:52 DEBUG manager.SqlManager: No connection paramenters specified. Using regular
API for making connection.
12/10/26 17:39:53 INFO tool.CodeGenTool: Beginning code generation
12/10/26 17:39:53 DEBUG manager.SqlManager: Using fetchSize for next query: 1000
12/10/26 17:39:53 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM [TABLENAME]
AS t WHERE 1=0
12/10/26 17:39:53 DEBUG orm.ClassWriter: selected columns:
12/10/26 17:39:53 DEBUG orm.ClassWriter:   Keyword
12/10/26 17:39:53 DEBUG orm.ClassWriter:   Item
12/10/26 17:39:53 DEBUG orm.ClassWriter:   Description
12/10/26 17:39:53 DEBUG orm.ClassWriter: Writing source file: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.java
12/10/26 17:39:53 DEBUG orm.ClassWriter: Table name: TABLENAME
12/10/26 17:39:53 DEBUG orm.ClassWriter: Columns: Keyword:-9, Item:-9, Description:-9,
12/10/26 17:39:53 DEBUG orm.ClassWriter: sourceFilename is TABLENAME.java
12/10/26 17:39:53 DEBUG orm.CompilationManager: Found existing /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/
12/10/26 17:39:53 INFO orm.CompilationManager: HADOOP_HOME is /usr/local/hadoop/libexec/..
12/10/26 17:39:53 DEBUG orm.CompilationManager: Adding source file: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.java
12/10/26 17:39:53 DEBUG orm.CompilationManager: Invoking javac with args:
12/10/26 17:39:53 DEBUG orm.CompilationManager:   -sourcepath
12/10/26 17:39:53 DEBUG orm.CompilationManager:   /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/
12/10/26 17:39:53 DEBUG orm.CompilationManager:   -d
12/10/26 17:39:53 DEBUG orm.CompilationManager:   /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/
12/10/26 17:39:53 DEBUG orm.CompilationManager:   -classpath
12/10/26 17:39:53 DEBUG orm.CompilationManager:   /usr/local/hadoop/libexec/../conf:/usr/lib/jvm/java-sun/lib/tools.jar:/usr/local/hadoop/libexec/..:/usr/local/hadoop/libexec/../hadoop-core-1.0.3.jar:/usr/local/hadoop/libexec/../lib/asm-3.2.jar:/usr/local/hadoop/libexec/../lib/aspectjrt-1.6.5.jar:/usr/local/hadoop/libexec/../lib/aspectjtools-1.6.5.jar:/usr/local/hadoop/libexec/../lib/avro-1.5.3.jar:/usr/local/hadoop/libexec/../lib/avro-ipc-1.5.3.jar:/usr/local/hadoop/libexec/../lib/avro-mapred-1.5.3.jar:/usr/local/hadoop/libexec/../lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop/libexec/../lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop/libexec/../lib/commons-cli-1.2.jar:/usr/local/hadoop/libexec/../lib/commons-codec-1.4.jar:/usr/local/hadoop/libexec/../lib/commons-collections-3.2.1.jar:/usr/local/hadoop/libexec/../lib/commons-configuration-1.6.jar:/usr/local/hadoop/libexec/../lib/commons-daemon-1.0.1.jar:/usr/local/hadoop/libexec/../lib/commons-digester-1.8.jar:/usr/local/hadoop/libexec/../lib/commons-el-1.0.jar:/usr/local/hadoop/libexec/../lib/commons-httpclient-3.0.1.jar:/usr/local/hadoop/libexec/../lib/commons-io-2.1.jar:/usr/local/hadoop/libexec/../lib/commons-lang-2.4.jar:/usr/local/hadoop/libexec/../lib/commons-logging-1.1.1.jar:/usr/local/hadoop/libexec/../lib/commons-logging-api-1.0.4.jar:/usr/local/hadoop/libexec/../lib/commons-math-2.1.jar:/usr/local/hadoop/libexec/../lib/commons-net-1.4.1.jar:/usr/local/hadoop/libexec/../lib/core-3.1.1.jar:/usr/local/hadoop/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/usr/local/hadoop/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/usr/local/hadoop/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/usr/local/hadoop/libexec/../lib/hsqldb-1.8.0.10.jar:/usr/local/hadoop/libexec/../lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop/libexec/../lib/jasper-compiler-5.5.12.jar:/usr/local/hadoop/libexec/../lib/jasper-runtime-5.5.12.jar:/usr/local/hadoop/libexec/../lib/jdeb-0.8.jar:/usr/local/hadoop/libexec/../lib/jersey-core-1.8.jar:/usr/local/hadoop/libexec/../lib/jersey-json-1.8.jar:/usr/local/hadoop/libexec/../lib/jersey-server-1.8.jar:/usr/local/hadoop/libexec/../lib/jets3t-0.6.1.jar:/usr/local/hadoop/libexec/../lib/jetty-6.1.26.jar:/usr/local/hadoop/libexec/../lib/jetty-util-6.1.26.jar:/usr/local/hadoop/libexec/../lib/jsch-0.1.42.jar:/usr/local/hadoop/libexec/../lib/junit-4.5.jar:/usr/local/hadoop/libexec/../lib/kfs-0.2.2.jar:/usr/local/hadoop/libexec/../lib/log4j-1.2.15.jar:/usr/local/hadoop/libexec/../lib/mockito-all-1.8.5.jar:/usr/local/hadoop/libexec/../lib/oro-2.0.8.jar:/usr/local/hadoop/libexec/../lib/servlet-api-2.5-20081211.jar:/usr/local/hadoop/libexec/../lib/slf4j-api-1.4.3.jar:/usr/local/hadoop/libexec/../lib/slf4j-log4j12-1.4.3.jar:/usr/local/hadoop/libexec/../lib/sqljdbc4.jar:/usr/local/hadoop/libexec/../lib/sqoop-1.4.2.jar:/usr/local/hadoop/libexec/../lib/sqoop-sqlserver-1.0.jar:/usr/local/hadoop/libexec/../lib/xmlenc-0.52.jar:/usr/local/hadoop/libexec/../lib/jsp-2.1/jsp-2.1.jar:/usr/local/hadoop/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/usr/local/hive/lib/hive_contrib.jar:/usr/local/hive/lib/commons-collections-3.2.1.jar:/usr/local/hive/lib/commons-codec-1.4.jar:/usr/local/hive/lib/xz-1.0.jar:/usr/local/hive/lib/stringtemplate-3.1-b1.jar:/usr/local/hive/lib/hbase-0.92.0.jar:/usr/local/hive/lib/commons-dbcp-1.4.jar:/usr/local/hive/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hive/lib/hive-contrib-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/hive-metastore-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/servlet-api-2.5-20081211.jar:/usr/local/hive/lib/antlr-runtime-3.0.1.jar:/usr/local/hive/lib/avro-mapred-1.7.1.jar:/usr/local/hive/lib/datanucleus-core-2.0.3.jar:/usr/local/hive/lib/datanucleus-enhancer-2.0.3.jar:/usr/local/hive/lib/hive-jdbc-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/commons-logging-1.0.4.jar:/usr/local/hive/lib/hive-serde-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/log4j-1.2.16.jar:/usr/local/hive/lib/avro-1.7.1.jar:/usr/local/hive/lib/commons-pool-1.5.4.jar:/usr/local/hive/lib/hive-hwi-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/datanucleus-rdbms-2.0.3.jar:/usr/local/hive/lib/hive-builtins-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/sqlline-1_0_2.jar:/usr/local/hive/lib/hive-pdk-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/libthrift-0.7.0.jar:/usr/local/hive/lib/jdo2-api-2.3-ec.jar:/usr/local/hive/lib/jline-0.9.94.jar:/usr/local/hive/lib/hive-service-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/json-20090211.jar:/usr/local/hive/lib/jetty-util-6.1.26.jar:/usr/local/hive/lib/derby-10.4.2.0.jar:/usr/local/hive/lib/commons-configuration-1.6.jar:/usr/local/hive/lib/jetty-6.1.26.jar:/usr/local/hive/lib/hive-exec-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/slf4j-log4j12-1.6.1.jar:/usr/local/hive/lib/commons-compress-1.4.1.jar:/usr/local/hive/lib/commons-logging-api-1.0.4.jar:/usr/local/hive/lib/hive-cli-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/hive-shims-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/jackson-xc-1.8.8.jar:/usr/local/hive/lib/hive-common-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/commons-lang-2.4.jar:/usr/local/hive/lib/datanucleus-connectionpool-2.0.3.jar:/usr/local/hive/lib/hive-hbase-handler-0.10.0-SNAPSHOT.jar:/usr/local/hive/lib/JavaEWAH-0.3.2.jar:/usr/local/hive/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hive/lib/antlr-2.7.7.jar:/usr/local/hive/lib/libfb303-0.7.0.jar:/usr/local/hive/lib/guava-11.0.2.jar:/usr/local/hive/lib/hbase-0.92.0-tests.jar:/usr/local/hive/lib/slf4j-api-1.6.1.jar:/usr/local/hive/lib/zookeeper-3.4.3.jar:/usr/local/hive/lib/commons-cli-1.2.jar:/usr/local/hive/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop/hadoop-core-1.0.3.jar:/usr/local/hadoop/lib/sqoop-1.4.2.jar
Note: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.java uses or overrides
a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
12/10/26 17:39:58 DEBUG orm.CompilationManager: Could not rename /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.java
to /home/xxxxx/./TABLENAME.java
org.apache.commons.io.FileExistsException: Destination '/home/xxxxx/./TABLENAME.java' already
exists
        at org.apache.commons.io.FileUtils.moveFile(FileUtils.java:2378)
        at org.apache.sqoop.orm.CompilationManager.compile(CompilationManager.java:227)
        at org.apache.sqoop.tool.CodeGenTool.generateORM(CodeGenTool.java:83)
        at org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:390)
        at org.apache.sqoop.tool.ImportAllTablesTool.run(ImportAllTablesTool.java:64)
        at org.apache.sqoop.Sqoop.run(Sqoop.java:145)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
        at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)
        at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)
        at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)
        at org.apache.sqoop.Sqoop.main(Sqoop.java:238)
        at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)
12/10/26 17:39:58 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.jar
12/10/26 17:39:58 DEBUG orm.CompilationManager: Scanning for .class files in directory: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946
12/10/26 17:39:58 DEBUG orm.CompilationManager: Got classfile: /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.class
-> TABLENAME.class
12/10/26 17:39:58 DEBUG orm.CompilationManager: Finished writing jar file /tmp/sqoop-xxxxx/compile/3a8443bba92a6cc65a6eef5b0999e946/TABLENAME.jar
12/10/26 17:39:58 DEBUG manager.CatalogQueryManager: Retrieving primary key for table 'TABLENAME'
with query SELECT kcu.COLUMN_NAME FROM INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc,   INFORMATION_SCHEMA.KEY_COLUMN_USAGE
kcu WHERE tc.TABLE_SCHEMA = kcu.TABLE_SCHEMA   AND tc.TABLE_NAME = kcu.TABLE_NAME   AND tc.CONSTRAINT_SCHEMA
= kcu.CONSTRAINT_SCHEMA   AND tc.CONSTRAINT_NAME = kcu.CONSTRAINT_NAME   AND tc.TABLE_SCHEMA
= (SELECT SCHEMA_NAME())   AND tc.TABLE_NAME = 'TABLENAME'   AND tc.CONSTRAINT_TYPE = 'PRIMARY
KEY'

12/10/26 17:39:58 WARN manager.CatalogQueryManager: The table TABLENAME contains a multi-column
primary key. Sqoop will default to the column Item only for this job.
12/10/26 17:39:58 DEBUG manager.CatalogQueryManager: Retrieving primary key for table 'TABLENAME'
with query SELECT kcu.COLUMN_NAME FROM INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc,   INFORMATION_SCHEMA.KEY_COLUMN_USAGE
kcu WHERE tc.TABLE_SCHEMA = kcu.TABLE_SCHEMA   AND tc.TABLE_NAME = kcu.TABLE_NAME   AND tc.CONSTRAINT_SCHEMA
= kcu.CONSTRAINT_SCHEMA   AND tc.CONSTRAINT_NAME = kcu.CONSTRAINT_NAME   AND tc.TABLE_SCHEMA
= (SELECT SCHEMA_NAME())   AND tc.TABLE_NAME = 'TABLENAME'   AND tc.CONSTRAINT_TYPE = 'PRIMARY
KEY'

12/10/26 17:39:58 WARN manager.CatalogQueryManager: The table TABLENAME contains a multi-column
primary key. Sqoop will default to the column Item only for this job.
12/10/26 17:39:58 INFO mapreduce.ImportJobBase: Beginning import of TABLENAME
12/10/26 17:39:58 DEBUG mapreduce.DataDrivenImportJob: Using table class: TABLENAME
12/10/26 17:39:58 DEBUG mapreduce.DataDrivenImportJob: Using InputFormat: class com.cloudera.sqoop.mapreduce.db.DataDrivenDBInputFormat
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/hadoop/lib/sqoop-1.4.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/hadoop/lib/sqljdbc4.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/hadoop/lib/sqoop-1.4.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/hadoop/lib/sqoop-1.4.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/avro-1.5.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/avro-ipc-1.5.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/commons-io-1.4.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/hsqldb-1.8.0.10.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/snappy-java-1.0.3.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/jopt-simple-3.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/sqljdbc4.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/avro-mapred-1.5.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/sqoop-sqlserver-1.0.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/jackson-core-asl-1.7.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/paranamer-2.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/jackson-mapper-asl-1.7.3.jar
12/10/26 17:39:58 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/local/sqoop/lib/ant-contrib-1.0b3.jar
12/10/26 17:40:02 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT MIN([Item]),
MAX([Item]) FROM [TABLENAME]
12/10/26 17:40:02 INFO mapred.JobClient: Cleaning up the staging area hdfs://SERVERNAME1:54310/user/xxxxx/.staging/job_201210261704_0005
12/10/26 17:40:02 ERROR security.UserGroupInformation: PriviledgedActionException as:xxxxx
cause:java.io.IOException: Unknown SQL data type: -9
12/10/26 17:40:02 ERROR tool.ImportAllTablesTool: Encountered IOException running import job:
java.io.IOException: Unknown SQL data type: -9



[xxxxx@servername ~]$ java -version
java version "1.6.0_33"
Java(TM) SE Runtime Environment (build 1.6.0_33-b03)
Java HotSpot(TM) 64-Bit Server VM (build 20.8-b03, mixed mode)

/usr/lib/jvm/jre-1.6.0-sun.x86_64/bin/java

-----Original Message-----
From: Jarek Jarcec Cecho [mailto:jarcec@apache.org] 
Sent: Friday, October 26, 2012 5:17 PM
To: user@sqoop.apache.org
Subject: Re: SQOOP 1.3.0 questions using SQL Server

Hi Artem,
yes nvarchar (sql type -9) should be supported on 1.4.2. Would you mind share with us entire
Sqoop command line and appropriate log generated with --verbose argument? Also what version
of Java are using?

Jarcec

On Fri, Oct 26, 2012 at 09:09:56PM +0000, Artem Ervits wrote:
> I have Sqoop 1.4.2 installed
> 
> Running sqoop version returns
> 
> Sqoop 1.4.2
> git commit id
> Compiled by ag on Tue Aug 14 17:37:19 IST 2012
> 
> When I tried to run a sqoop import-all-tables I got "Unknown SQL Data type: -9". From
the discussion below, nvarchar should be supported in this version of sqoop, please advise?
> 
> Thank you.
> 
> -----Original Message-----
> From: Jarek Jarcec Cecho [mailto:jarcec@apache.org]
> Sent: Monday, October 15, 2012 4:34 PM
> To: user@sqoop.apache.org
> Cc: Marc Sturm
> Subject: Re: SQOOP 1.3.0 questions using SQL Server
> 
> Hi Marc,
> please find my answers inline:
> 
> On Mon, Oct 15, 2012 at 07:25:28PM +0000, Artem Ervits wrote:
> > I installed SQOOP 1.4.2 and was able to compress and then uncompress the data. I
used the bin.tar.gz installer.
> 
> Great!
> 
> > 
> > 1.       I did try to use the git repo to compile the sqoop jar and was able to
compile it no problem which resulted in 1.4.3 snapshot jar using command ant jar-all. I noticed
that  all other jars like avro are not present in the build or lib directory. How may I go
about downloading source and compiling it to include avro, etc?
> 
> If you need to build entire distribution (like the one that is offered in download page)
including (almost) all dependency jars you need to run command "ant -Dhadoopversion=$hadoopVersion
tar", where $hadoopVersion is your hadoop version. Supported values are 20, 23, 100 and 200.
> 
> > 2.       I also tried to use the incremental switch, is this a correct command:
"sqoop import -incremental 'lastmodified'" or does the command look some other way?
> 
> Yes that is the intended usage. You might find more information about incremental imports
in Sqoop documentation:
> 
> http://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_incremental_im
> ports
> 
> > Thank you.
> 
> Jarcec
> 
> > 
> > 
> > From: Artem Ervits [mailto:are9004@nyp.org]
> > Sent: Friday, October 12, 2012 5:17 PM
> > To: user@sqoop.apache.org
> > Cc: Marc Sturm
> > Subject: RE: SQOOP 1.3.0 questions using SQL Server
> > 
> > Even better!
> > 
> > Thank you.
> > 
> > From: Chalcy
> > [mailto:chalcy@gmail.com]<mailto:[mailto:chalcy@gmail.com]>
> > Sent: Friday, October 12, 2012 4:18 PM
> > To: user@sqoop.apache.org<mailto:user@sqoop.apache.org>
> > Subject: Re: SQOOP 1.3.0 questions using SQL Server
> > 
> > Sqoop 1.4.3 freshly built even works very well with sql server. Just 
> > get the sqoop jar and replace it and should work just fine.  I so 
> > far never had any issues changing sqoop version :)
> > 
> > Good luck!
> > 
> > Thanks,
> > Chalcy
> > On Fri, Oct 12, 2012 at 2:16 PM, Artem Ervits <are9004@nyp.org<mailto:are9004@nyp.org>>
wrote:
> > Actually I was just able to test 3) below and it succeeded, I cast the nvarchar
as char.
> > 
> > Thank you again, I will do 1) and 2) and report back the results.
> > 
> > -----Original Message-----
> > From: Jarek Jarcec Cecho
> > [mailto:jarcec@apache.org<mailto:jarcec@apache.org>]
> > Sent: Friday, October 12, 2012 1:51 PM
> > To: user@sqoop.apache.org<mailto:user@sqoop.apache.org>
> > Subject: Re: SQOOP 1.3.0 questions using SQL Server Hi Artem, Sqoop
> > 1.3.0 is very old release, so I would definitely recommend to upgrade to latest
1.4.2. Please note that this release include SQOOP-480[1] that fixes some issues for connectors
compiled against 1.3 in 1.4. Sqoop is simple command line utility so you might try to download
release into some temporary folder and verify that everything is working prior upgrading Sqoop
system wise.
> > 
> > Specifically to your questions
> > 
> > 1) --incremental
> > Usually Sqoop will print out error message first followed by help page, so I would
advise checking few initial lines to see what exactly is wrong.
> > 
> > 2) --compress
> > Please try to reproduce this issue on 1.4.2 and file jira if you'll be able to reproduce
it.
> > 
> > 3) --as-avrodatafile
> > Based on [2] SQL type -9 corresponds to data type NVARCHAR. We've added support
for NVARCHAR in SQOOP-323 [3] that is part of Sqoop since version 1.4.0-incubating - e.g.
1.3.0 version do not support NVARCHAR type out of the box.
> > 
> > Jarcec
> > 
> > Links:
> > 1: https://issues.apache.org/jira/browse/SQOOP-480
> > 2: 
> > http://docs.oracle.com/javase/6/docs/api/constant-values.html#java.s
> > ql
> > .Types.DECIMAL
> > 3: https://issues.apache.org/jira/browse/SQOOP-323
> > 
> > On Fri, Oct 12, 2012 at 05:29:41PM +0000, Artem Ervits wrote:
> > >
> > > Hello all,
> > >
> > > I'm testing a few switches for Sqoop import and I'm having the following problems.
> > >
> > > When I use the -incremental switch, the command fails and shows me the help
page.
> > > When I use the -compress switch, the command works but when I try to uncompress
the results, it says gzip undefined compression code. I also tried to explicitly state the
codec and it would still not append the compression extension to the files nor am I able to
uncompress the data.
> > > When I use the -as-avrodatafile I get ERROR tool.ImportTool: 
> > > Imported Failed: Cannot convert SQL type -9
> > >
> > > Any ideas? I am not sure if upgrading Sqoop will fix it because SQL Server
connector specifically required a 1.3.0 release.
> > >
> > > Thanks.
> > >
> > > Artem Ervits
> > > Data Analyst
> > > New York Presbyterian Hospital
> > >
> > >
> > >
> > > --------------------
> > >
> > > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > >
> > >
> > >
> > >
> > > --------------------
> > >
> > > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > >
> > >
> > >
> > 
> > 
> > --------------------
> > 
> > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > 
> > 
> > 
> > 
> > --------------------
> > 
> > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > 
> > 
> > 
> > 
> > 
> > --------------------
> > 
> > Confidential Information subject to NYP's (and its affiliates') information management
and security policies (http://infonet.nyp.org/QA/HospManual/).
> > 
> > 
> > 
> > --------------------
> > 
> > 
> > 
> > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > 
> > 
> > 
> > 
> > 
> > ________________________________
> > 
> > Confidential Information subject to NYP's (and its affiliates') information management
and security policies (http://infonet.nyp.org/QA/HospitalManual).
> > 
> > 
> > --------------------
> > Confidential Information subject to NYP's (and its affiliates') information management
and security policies (http://infonet.nyp.org/QA/HospManual/).
> > 
> > 
> > 
> > --------------------
> > 
> > This electronic message is intended to be for the use only of the named recipient,
and may contain information that is confidential or privileged.  If you are not the intended
recipient, you are hereby notified that any disclosure, copying, distribution or use of the
contents of this message is strictly prohibited.  If you have received this message in error
or are not the named recipient, please notify us immediately by contacting the sender at the
electronic mail address noted above, and delete and destroy all copies of this message.  Thank
you.
> > 
> > 
> > 
> 
> 
> --------------------
> 
> This electronic message is intended to be for the use only of the named recipient, and
may contain information that is confidential or privileged.  If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution or use of the contents
of this message is strictly prohibited.  If you have received this message in error or are
not the named recipient, please notify us immediately by contacting the sender at the electronic
mail address noted above, and delete and destroy all copies of this message.  Thank you.
> 
> 
> 
> 
> --------------------
> 
> This electronic message is intended to be for the use only of the named recipient, and
may contain information that is confidential or privileged.  If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution or use of the contents
of this message is strictly prohibited.  If you have received this message in error or are
not the named recipient, please notify us immediately by contacting the sender at the electronic
mail address noted above, and delete and destroy all copies of this message.  Thank you.
> 
> 
> 


--------------------

This electronic message is intended to be for the use only of the named recipient, and may
contain information that is confidential or privileged.  If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution or use of the contents
of this message is strictly prohibited.  If you have received this message in error or are
not the named recipient, please notify us immediately by contacting the sender at the electronic
mail address noted above, and delete and destroy all copies of this message.  Thank you.




--------------------

This electronic message is intended to be for the use only of the named recipient, and may
contain information that is confidential or privileged.  If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution or use of the contents
of this message is strictly prohibited.  If you have received this message in error or are
not the named recipient, please notify us immediately by contacting the sender at the electronic
mail address noted above, and delete and destroy all copies of this message.  Thank you.




Mime
View raw message