kylin-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 周湘伦 <jayzho...@gmail.com>
Subject Re: Some error happened when i build cube
Date Fri, 09 Jun 2017 06:27:27 GMT
HI,ShaoFeng

I had replaced hbase-server-1.1.1.jar which in hive/lib
to hbase-server-1.2.5.jar.
The above problem has been solved.
Thank you!

But another problem arises:

14:16:52.441 [Job 90064ee6-2b3a-4ef7-b035-33f40391aafb-141] ERROR
org.apache.kylin.job.execution.AbstractExecutable - error running
Executable: MapReduceExecutable{id=90064ee6-2b3a-4ef7-b035-33f40391aafb-15,
name=Convert Cuboid Data to HFile, state=RUNNING}
14:16:52.496 [Job 90064ee6-2b3a-4ef7-b035-33f40391aafb-141] ERROR
org.apache.kylin.job.execution.AbstractExecutable - error running
Executable: CubingJob{id=90064ee6-2b3a-4ef7-b035-33f40391aafb,
name=Kylin_Sample_cube_1 - 20120101000000_20130101000000 - BUILD -
GMT+08:00 2017-06-08 15:06:49, state=RUNNING}
14:16:52.537 [pool-9-thread-1] ERROR
org.apache.kylin.job.impl.threadpool.DefaultScheduler - ExecuteException
job:90064ee6-2b3a-4ef7-b035-33f40391aafb
org.apache.kylin.job.exception.ExecuteException:
org.apache.kylin.job.exception.ExecuteException:
java.lang.NoSuchFieldError: DEFAULT_TEMPORARY_HDFS_DIRECTORY
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:134)
~[kylin-core-job-2.0.0.jar:2.0.0]
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:142)
[kylin-core-job-2.0.0.jar:2.0.0]
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[?:1.8.0_131]
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[?:1.8.0_131]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_131]
Caused by: org.apache.kylin.job.exception.ExecuteException:
java.lang.NoSuchFieldError: DEFAULT_TEMPORARY_HDFS_DIRECTORY
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:134)
~[kylin-core-job-2.0.0.jar:2.0.0]
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:64)
~[kylin-core-job-2.0.0.jar:2.0.0]
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
~[kylin-core-job-2.0.0.jar:2.0.0]
... 4 more


I had configured this option in hbase-site.xml:
<property>
        <name>hbase.fs.tmp.dir</name>
        <value>hdfs://master:9000/hbase/tmp/hadoop-staging</value>
      <!-- <value>/tmp/hadoop-staging</value>-->
        <description>i have tried these configurations</description>
    </property>

I had even configured this option in kylin_job_conf.xml.
But there is no effect.

Of course i have created director in local file system and hdfs.

Can you tell me how to solve this problem?
Than you!

2017-06-05 13:04 GMT+08:00 ShaoFeng Shi <shaofengshi@apache.org>:

> The method "getChecksumObject" does not exist in "
> org.apache.hadoop.hbase.util.ChecksumType" in hbase 1.2.5. While in your
> environment the "org.apache.hadoop.hbase.io.hfile.ChecksumUtil" still
> invoking to it. It indicates there is a "hbase-server*.jar" which isn't
> v1.2.5. Please search your environment globally to identify and remove it.
>
> Usually we suggest using a formal release of CDH/HDP/MapR, which won't have
> such version conflict. If you install these components separately, there
> will be such environment issues.
>
> 2017-06-03 18:29 GMT+08:00 周湘伦 <jayzhouxl@gmail.com>:
>
> > Hi,all
> > When i built cube,an error occurred in step 16(Convert Cuboid Data to
> > HFile).
> >
> > The version as belows:
> > hadoop-2.8.0,hbase-1.2.5,jdk1.8.0_131,kylin-2.0.0
> >
> > The error logs in hadoop/userlogs,the log is shown below:
> >
> > 2017-06-03 17:57:04,106 FATAL [main] org.apache.hadoop.mapred.YarnChild:
> > Error running child : java.lang.NoSuchMethodError:
> > org.apache.hadoop.hbase.util.ChecksumType.getChecksumObject()Ljava/util/
> > zip/Checksum;
> >         at
> > org.apache.hadoop.hbase.io.hfile.ChecksumUtil.generateChecksums(
> > ChecksumUtil.java:73)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.
> > finishBlock(HFileBlock.java:943)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.
> > ensureBlockReady(HFileBlock.java:895)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.
> > finishBlockAndWriteHeaderAndData(HFileBlock.java:1011)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.
> > writeHeaderAndData(HFileBlock.java:997)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexWriter.
> > writeIndexBlocks(HFileBlockIndex.java:883)
> >         at
> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.close(
> > HFileWriterV2.java:331)
> >         at
> > org.apache.hadoop.hbase.regionserver.StoreFile$Writer.
> > close(StoreFile.java:996)
> >         at
> > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$
> > 1.close(HFileOutputFormat2.java:269)
> >         at
> > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$
> > 1.close(HFileOutputFormat2.java:277)
> >         at
> > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.close(
> > ReduceTask.java:550)
> >         at
> > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:629)
> >         at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
> >         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:175)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> > org.apache.hadoop.security.UserGroupInformation.doAs(
> > UserGroupInformation.java:1807)
> >         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:169)
> >
> > According to others' suggest,i had copy jars which in hadoop-2.8.0 to
> > hbase-1.2.5,but the problem is still not resolved.
> >
> > How can we solve the problem?
> >
> > Thanks a lot.
> >
>
>
>
> --
> Best regards,
>
> Shaofeng Shi 史少锋
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message