hive-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: Hive-trunk-h0.21 #903
Date Fri, 19 Aug 2011 03:02:52 GMT
See <https://builds.apache.org/job/Hive-trunk-h0.21/903/changes>

Changes:

[cws] HIVE-2315. DatabaseMetadata.getColumns() does not return partition column names for
a table (Patrick Hunt via cws)

------------------------------------------
[...truncated 32096 lines...]
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select * from testhivedrivertable limit 10
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-35_784_1014922431364815295/-mr-10000
    [junit] POSTHOOK: query: select * from testhivedrivertable limit 10
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-35_784_1014922431364815295/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_566137889.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (num int)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (num int)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_923144116.txt>
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_1448946765.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] Copying data from <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table default.testhivedrivertable
    [junit] POSTHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-37_051_4640888108895700952/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-18 19:47:39,748 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] Execution completed successfully
    [junit] Mapred Local Task Succeeded . Convert the Join into MapJoin
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-37_051_4640888108895700952/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-39_901_386971226042768968/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-18 19:47:42,604 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] Execution completed successfully
    [junit] Mapred Local Task Succeeded . Convert the Join into MapJoin
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-39_901_386971226042768968/-mr-10000
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-42_842_5117998180441020353/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-18 19:47:45,530 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] Execution completed successfully
    [junit] Mapred Local Task Succeeded . Convert the Join into MapJoin
    [junit] POSTHOOK: query: select key, value from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-42_842_5117998180441020353/-mr-10000
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_1806758410.txt>
    [junit] PREHOOK: query: drop table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] POSTHOOK: query: drop table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] PREHOOK: type: DROPTABLE
    [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string)
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Output: default@testhivedrivertable
    [junit] Copying data from <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt>
    [junit] Loading data to table default.testhivedrivertable
    [junit] POSTHOOK: query: load data local inpath '<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Output: default@testhivedrivertable
    [junit] OK
    [junit] PREHOOK: query: select key, value from testhivedrivertable where key > 10
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-46_448_616033827090499900/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks is set to 0 since there's no reduce operator
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-18 19:47:49,222 null map = 100%,  reduce = 0%
    [junit] Ended Job = job_local_0001
    [junit] Execution completed successfully
    [junit] Mapred Local Task Succeeded . Convert the Join into MapJoin
    [junit] POSTHOOK: query: select key, value from testhivedrivertable where key > 10
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-46_448_616033827090499900/-mr-10000
    [junit] OK
    [junit] SLF4J: Class path contains multiple SLF4J bindings.
    [junit] SLF4J: Found binding in [jar:<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/ivy/lib/default/slf4j-log4j12-1.6.1-javadoc.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    [junit] SLF4J: Found binding in [jar:<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/ivy/lib/default/slf4j-log4j12-1.6.1-sources.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    [junit] SLF4J: Found binding in [jar:<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/ivy/lib/default/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    [junit] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    [junit] PREHOOK: query: select count(1) as c from testhivedrivertable
    [junit] PREHOOK: type: DROPTABLE
    [junit] PREHOOK: Input: default@testhivedrivertable
    [junit] PREHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-49_457_1683684205424652127/-mr-10000
    [junit] Total MapReduce jobs = 1
    [junit] Launching Job 1 out of 1
    [junit] Number of reduce tasks determined at compile time: 1
    [junit] In order to change the average load for a reducer (in bytes):
    [junit]   set hive.exec.reducers.bytes.per.reducer=<number>
    [junit] In order to limit the maximum number of reducers:
    [junit]   set hive.exec.reducers.max=<number>
    [junit] In order to set a constant number of reducers:
    [junit]   set mapred.reduce.tasks=<number>
    [junit] Job running in-process (local Hadoop)
    [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0
    [junit] 2011-08-18 19:47:52,287 null map = 100%,  reduce = 100%
    [junit] Ended Job = job_local_0001
    [junit] Execution completed successfully
    [junit] Mapred Local Task Succeeded . Convert the Join into MapJoin
    [junit] POSTHOOK: query: select count(1) as c from testhivedrivertable
    [junit] POSTHOOK: type: DROPTABLE
    [junit] POSTHOOK: Input: default@testhivedrivertable
    [junit] POSTHOOK: Output: file:/tmp/jenkins/hive_2011-08-18_19-47-49_457_1683684205424652127/-mr-10000
    [junit] OK
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_1967745850.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.jar>
does not exist
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_139144969.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.txt>
does not exist
    [junit] Hive history file=<https://builds.apache.org/job/Hive-trunk-h0.21/903/artifact/hive/build/service/tmp/hive_job_log_jenkins_201108181947_1724198390.txt>
    [junit] <https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/service/../data/files/invalidpath/sample.zip>
does not exist
    [junit] ------------- ---------------- ---------------
    [junit] 
    [junit] Testcase: testExecute took 9.818 sec
    [junit] Testcase: testNonHiveCommand took 0.98 sec
    [junit] Testcase: testMetastore took 0.267 sec
    [junit] Testcase: testGetClusterStatus took 0.068 sec
    [junit] Testcase: testFetch took 9.311 sec
    [junit] Testcase: testDynamicSerde took 6.752 sec
    [junit] Testcase: testAddJarShouldFailIfJarNotExist took 0.043 sec
    [junit] Testcase: testAddFileShouldFailIfFileNotExist took 0.043 sec
    [junit] Testcase: testAddArchiveShouldFailIfFileNotExist took 0.042 sec

BUILD FAILED
<https://builds.apache.org/job/Hive-trunk-h0.21/ws/hive/build.xml>:208: Keepgoing execution:
1 of 10 iterations failed.

Total time: 285 minutes 40 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Updating HIVE-2315
Recording test results
ERROR: Failed to archive test reports
hudson.util.IOException2: remote file operation failed: <https://builds.apache.org/job/Hive-trunk-h0.21/ws/>
at hudson.remoting.Channel@56473b73:ubuntu4
	at hudson.FilePath.act(FilePath.java:754)
	at hudson.FilePath.act(FilePath.java:740)
	at hudson.tasks.junit.JUnitParser.parse(JUnitParser.java:83)
	at hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:123)
	at hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:135)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
	at hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:682)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:657)
	at hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:635)
	at hudson.model.Build$RunnerImpl.post2(Build.java:162)
	at hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:604)
	at hudson.model.Run.run(Run.java:1401)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:230)
Caused by: java.io.IOException: Remote call on ubuntu4 failed
	at hudson.remoting.Channel.call(Channel.java:677)
	at hudson.FilePath.act(FilePath.java:747)
	... 14 more
Caused by: java.lang.NoClassDefFoundError: Could not initialize class hudson.tasks.junit.TestResult
	at hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:114)
	at hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:87)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1994)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:287)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
	at java.lang.Thread.run(Thread.java:636)


Mime
View raw message