hadoop-general mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Konstantin Shvachko <shv.had...@gmail.com>
Subject Re: Hadoop-common-trunk-Commit is failing since 01/19/2011
Date Tue, 01 Feb 2011 00:27:34 GMT
Giri
looks like the last run you started failed the same way as previous ones.
Any thoughts on what's going on?
Thanks,
--Konstantin

On Mon, Jan 31, 2011 at 3:33 PM, Giridharan Kesavan
<gkesavan@yahoo-inc.com>wrote:

> ant mvn-deploy would publish snapshot artifact to the apache maven
> repository as long you have the right credentials in ~/.m2/settings.xml.
>
> For settings.xml template pls look at
> http://wiki.apache.org/hadoop/HowToRelease
>
> I'm pushing the latest common artifacts now.
>
> -Giri
>
>
>
> On Jan 31, 2011, at 3:11 PM, Jakob Homan wrote:
>
> > By manually installing a new core jar into the cache, I can compile
> > trunk.  Looks like we just need to kick a new Core into maven.  Are
> > there instructions somewhere for committers to do this?  I know Nigel
> > and Owen know how, but I don't know if the knowledge is diffused past
> > them.
> > -Jakob
> >
> >
> > On Mon, Jan 31, 2011 at 1:57 PM, Konstantin Shvachko
> > <shv.hadoop@gmail.com> wrote:
> >> Current trunk for HDFS and MapReduce are not compiling at the moment.
> Try to
> >> build trunk.
> >> This is the result of that changes to common api introduced by
> HADOOP-6904
> >> are not promoted to HDFS and MR trunks.
> >> HDFS-1335 and MAPREDUCE-2263 depend on these changes.
> >>
> >> Common is not promoted to HDFS and MR because Hadoop-Common-trunk-Commit
> >> build is broken. See here.
> >>
> https://hudson.apache.org/hudson/view/G-L/view/Hadoop/job/Hadoop-Common-trunk-Commit/
> >>
> >> As I see the last successful build was on 01/19, which integrated
> >> HADOOP-6864.
> >> I think this is when JNI changes were introduced, which cannot be
> digested
> >> by Hudson since then.
> >>
> >> Anybody with gcc active could you please verify if the problem is caused
> by
> >> HADOOP-6864.
> >>
> >> Thanks,
> >> --Konstantin
> >>
> >> On Mon, Jan 31, 2011 at 1:36 PM, Ted Dunning <tdunning@maprtech.com>
> wrote:
> >>
> >>> The has been a problem with more than one build failing (Mahout is the
> one
> >>> that I saw first) due to a change in maven version which meant that the
> >>> clover license isn't being found properly.  At least, that is the tale
> I
> >>> heard from infra.
> >>>
> >>> On Mon, Jan 31, 2011 at 1:31 PM, Eli Collins <eli@cloudera.com> wrote:
> >>>
> >>>> Hey Konstantin,
> >>>>
> >>>> The only build breakage I saw from HADOOP-6904 is MAPREDUCE-2290,
> >>>> which was fixed.  Trees from trunk are compiling against each other
> >>>> for me (eg each installed to a local maven repo), perhaps the upstream
> >>>> maven repo hasn't been updated with the latest bits yet.
> >>>>
> >>>> Thanks,
> >>>> Eli
> >>>>
> >>>> On Mon, Jan 31, 2011 at 12:14 PM, Konstantin Shvachko
> >>>> <shv.hadoop@gmail.com> wrote:
> >>>>> Sending this to general to attract urgent attention.
> >>>>> Both HDFS and MapReduce are not compiling since
> >>>>> HADOOP-6904 and its hdfs and MP counterparts were committed.
> >>>>> The problem is not with this patch as described below, but I think
> >>> those
> >>>>> commits should be reversed if Common integration build cannot be
> >>>>> restored promptly.
> >>>>>
> >>>>> Thanks,
> >>>>> --Konstantin
> >>>>>
> >>>>>
> >>>>> On Fri, Jan 28, 2011 at 5:53 PM, Konstantin Shvachko
> >>>>> <shv.hadoop@gmail.com>wrote:
> >>>>>
> >>>>>> I see Hadoop-common-trunk-Commit is failing and not sending
any
> >>> emails.
> >>>>>> It times out on native compilation and aborts.
> >>>>>> Therefore changes are not integrated, and now it lead to hdfs
and
> >>>> mapreduce
> >>>>>> both not compiling.
> >>>>>> Can somebody please take a look at this.
> >>>>>> The last few lines of the build are below.
> >>>>>>
> >>>>>> Thanks
> >>>>>> --Konstantin
> >>>>>>
> >>>>>>     [javah] [Loaded
> >>>>
> >>>
> /grid/0/hudson/hudson-slave/workspace/Hadoop-Common-trunk-Commit/trunk/build/classes/org/apache/hadoop/security/JniBasedUnixGroupsMapping.class]
> >>>>>>
> >>>>>>     [javah] [Loaded
> >>>>
> >>>
> /homes/hudson/tools/java/jdk1.6.0_11-32/jre/lib/rt.jar(java/lang/Object.class)]
> >>>>>>     [javah] [Forcefully writing file
> >>>>
> >>>
> /grid/0/hudson/hudson-slave/workspace/Hadoop-Common-trunk-Commit/trunk/build/native/Linux-i386-32/src/org/apache/hadoop/security/org_apache_hadoop_security_JniBasedUnixGroupsNetgroupMapping.h]
> >>>>>>
> >>>>>>      [exec] checking for gcc... gcc
> >>>>>>      [exec] checking whether the C compiler works... yes
> >>>>>>      [exec] checking for C compiler default output file name...
> a.out
> >>>>>>      [exec] checking for suffix of executables...
> >>>>>>
> >>>>>> Build timed out. Aborting
> >>>>>> Build was aborted
> >>>>>> [FINDBUGS] Skipping publisher since build result is ABORTED
> >>>>>> Publishing Javadoc
> >>>>>> Archiving artifacts
> >>>>>> Recording test results
> >>>>>> No test report files were found. Configuration error?
> >>>>>>
> >>>>>> Recording fingerprints
> >>>>>>      [exec] Terminated
> >>>>>> Publishing Clover coverage report...
> >>>>>> No Clover report will be published due to a Build Failure
> >>>>>> No emails were triggered.
> >>>>>> Finished: ABORTED
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message