hadoop-general mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Konstantin Shvachko <shv.had...@gmail.com>
Subject Re: Hadoop-common-trunk-Commit is failing since 01/19/2011
Date Tue, 01 Feb 2011 19:44:33 GMT
Giri,
Thanks a lot for fixing this.
I see it is working now.
--Konstantin

On Tue, Feb 1, 2011 at 11:27 AM, Giridharan Kesavan
<gkesavan@yahoo-inc.com>wrote:

> Konstantin,
>
> trunk/artifacts gets populated when the jar and the tar ant target are
> successful.
>
> The main reason for the build failure so far is the build abort time
> configuration. It was set to 30mins.
> I have increased the build abort time and the builds are going on fine
>
> https://hudson.apache.org/hudson/view/G-L/view/Hadoop/job/Hadoop-Common-trunk-Commit
>
>
> Thanks,
> Giri
>
> On Feb 1, 2011, at 12:40 AM, Konstantin Shvachko wrote:
>
> > Giri,
> >
> > Looking at configuration of Hadoop-Common-trunk-Commit/
> > There seems to be errors in the Post-build Actions.
> > It is complaining that
> > 'trunk' exists but not 'trunk/artifacts/...'
> > Is it possible that this misconfiguration is the reason of failures?
> >
> > --Konstantin
> >
> >
> > On Mon, Jan 31, 2011 at 4:40 PM, Giridharan Kesavan
> > <gkesavan@yahoo-inc.com>wrote:
> >
> >> Konstantin,
> >>
> >> I think I need to restart the slave which is running the commit build.
> For
> >> now I have published the common artifact manually from commandline.
> >>
> >> Thanks,
> >> Giri
> >>
> >> On Jan 31, 2011, at 4:27 PM, Konstantin Shvachko wrote:
> >>
> >>> Giri
> >>> looks like the last run you started failed the same way as previous
> ones.
> >>> Any thoughts on what's going on?
> >>> Thanks,
> >>> --Konstantin
> >>>
> >>> On Mon, Jan 31, 2011 at 3:33 PM, Giridharan Kesavan
> >>> <gkesavan@yahoo-inc.com>wrote:
> >>>
> >>>> ant mvn-deploy would publish snapshot artifact to the apache maven
> >>>> repository as long you have the right credentials in
> ~/.m2/settings.xml.
> >>>>
> >>>> For settings.xml template pls look at
> >>>> http://wiki.apache.org/hadoop/HowToRelease
> >>>>
> >>>> I'm pushing the latest common artifacts now.
> >>>>
> >>>> -Giri
> >>>>
> >>>>
> >>>>
> >>>> On Jan 31, 2011, at 3:11 PM, Jakob Homan wrote:
> >>>>
> >>>>> By manually installing a new core jar into the cache, I can compile
> >>>>> trunk.  Looks like we just need to kick a new Core into maven. 
Are
> >>>>> there instructions somewhere for committers to do this?  I know
Nigel
> >>>>> and Owen know how, but I don't know if the knowledge is diffused
past
> >>>>> them.
> >>>>> -Jakob
> >>>>>
> >>>>>
> >>>>> On Mon, Jan 31, 2011 at 1:57 PM, Konstantin Shvachko
> >>>>> <shv.hadoop@gmail.com> wrote:
> >>>>>> Current trunk for HDFS and MapReduce are not compiling at the
> moment.
> >>>> Try to
> >>>>>> build trunk.
> >>>>>> This is the result of that changes to common api introduced
by
> >>>> HADOOP-6904
> >>>>>> are not promoted to HDFS and MR trunks.
> >>>>>> HDFS-1335 and MAPREDUCE-2263 depend on these changes.
> >>>>>>
> >>>>>> Common is not promoted to HDFS and MR because
> >> Hadoop-Common-trunk-Commit
> >>>>>> build is broken. See here.
> >>>>>>
> >>>>
> >>
> https://hudson.apache.org/hudson/view/G-L/view/Hadoop/job/Hadoop-Common-trunk-Commit/
> >>>>>>
> >>>>>> As I see the last successful build was on 01/19, which integrated
> >>>>>> HADOOP-6864.
> >>>>>> I think this is when JNI changes were introduced, which cannot
be
> >>>> digested
> >>>>>> by Hudson since then.
> >>>>>>
> >>>>>> Anybody with gcc active could you please verify if the problem
is
> >> caused
> >>>> by
> >>>>>> HADOOP-6864.
> >>>>>>
> >>>>>> Thanks,
> >>>>>> --Konstantin
> >>>>>>
> >>>>>> On Mon, Jan 31, 2011 at 1:36 PM, Ted Dunning <tdunning@maprtech.com
> >
> >>>> wrote:
> >>>>>>
> >>>>>>> The has been a problem with more than one build failing
(Mahout is
> >> the
> >>>> one
> >>>>>>> that I saw first) due to a change in maven version which
meant that
> >> the
> >>>>>>> clover license isn't being found properly.  At least, that
is the
> >> tale
> >>>> I
> >>>>>>> heard from infra.
> >>>>>>>
> >>>>>>> On Mon, Jan 31, 2011 at 1:31 PM, Eli Collins <eli@cloudera.com>
> >> wrote:
> >>>>>>>
> >>>>>>>> Hey Konstantin,
> >>>>>>>>
> >>>>>>>> The only build breakage I saw from HADOOP-6904 is MAPREDUCE-2290,
> >>>>>>>> which was fixed.  Trees from trunk are compiling against
each
> other
> >>>>>>>> for me (eg each installed to a local maven repo), perhaps
the
> >> upstream
> >>>>>>>> maven repo hasn't been updated with the latest bits
yet.
> >>>>>>>>
> >>>>>>>> Thanks,
> >>>>>>>> Eli
> >>>>>>>>
> >>>>>>>> On Mon, Jan 31, 2011 at 12:14 PM, Konstantin Shvachko
> >>>>>>>> <shv.hadoop@gmail.com> wrote:
> >>>>>>>>> Sending this to general to attract urgent attention.
> >>>>>>>>> Both HDFS and MapReduce are not compiling since
> >>>>>>>>> HADOOP-6904 and its hdfs and MP counterparts were
committed.
> >>>>>>>>> The problem is not with this patch as described
below, but I
> think
> >>>>>>> those
> >>>>>>>>> commits should be reversed if Common integration
build cannot be
> >>>>>>>>> restored promptly.
> >>>>>>>>>
> >>>>>>>>> Thanks,
> >>>>>>>>> --Konstantin
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On Fri, Jan 28, 2011 at 5:53 PM, Konstantin Shvachko
> >>>>>>>>> <shv.hadoop@gmail.com>wrote:
> >>>>>>>>>
> >>>>>>>>>> I see Hadoop-common-trunk-Commit is failing
and not sending any
> >>>>>>> emails.
> >>>>>>>>>> It times out on native compilation and aborts.
> >>>>>>>>>> Therefore changes are not integrated, and now
it lead to hdfs
> and
> >>>>>>>> mapreduce
> >>>>>>>>>> both not compiling.
> >>>>>>>>>> Can somebody please take a look at this.
> >>>>>>>>>> The last few lines of the build are below.
> >>>>>>>>>>
> >>>>>>>>>> Thanks
> >>>>>>>>>> --Konstantin
> >>>>>>>>>>
> >>>>>>>>>>   [javah] [Loaded
> >>>>>>>>
> >>>>>>>
> >>>>
> >>
> /grid/0/hudson/hudson-slave/workspace/Hadoop-Common-trunk-Commit/trunk/build/classes/org/apache/hadoop/security/JniBasedUnixGroupsMapping.class]
> >>>>>>>>>>
> >>>>>>>>>>   [javah] [Loaded
> >>>>>>>>
> >>>>>>>
> >>>>
> >>
> /homes/hudson/tools/java/jdk1.6.0_11-32/jre/lib/rt.jar(java/lang/Object.class)]
> >>>>>>>>>>   [javah] [Forcefully writing file
> >>>>>>>>
> >>>>>>>
> >>>>
> >>
> /grid/0/hudson/hudson-slave/workspace/Hadoop-Common-trunk-Commit/trunk/build/native/Linux-i386-32/src/org/apache/hadoop/security/org_apache_hadoop_security_JniBasedUnixGroupsNetgroupMapping.h]
> >>>>>>>>>>
> >>>>>>>>>>    [exec] checking for gcc... gcc
> >>>>>>>>>>    [exec] checking whether the C compiler works...
yes
> >>>>>>>>>>    [exec] checking for C compiler default output
file name...
> >>>> a.out
> >>>>>>>>>>    [exec] checking for suffix of executables...
> >>>>>>>>>>
> >>>>>>>>>> Build timed out. Aborting
> >>>>>>>>>> Build was aborted
> >>>>>>>>>> [FINDBUGS] Skipping publisher since build result
is ABORTED
> >>>>>>>>>> Publishing Javadoc
> >>>>>>>>>> Archiving artifacts
> >>>>>>>>>> Recording test results
> >>>>>>>>>> No test report files were found. Configuration
error?
> >>>>>>>>>>
> >>>>>>>>>> Recording fingerprints
> >>>>>>>>>>    [exec] Terminated
> >>>>>>>>>> Publishing Clover coverage report...
> >>>>>>>>>> No Clover report will be published due to a
Build Failure
> >>>>>>>>>> No emails were triggered.
> >>>>>>>>>> Finished: ABORTED
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>
> >>>>
> >>
> >>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message