incubator-bigtop-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From r..@apache.org
Subject svn commit: r1197368 - in /incubator/bigtop/branches/hadoop-0.23: ./ bigtop-packages/src/common/mahout/ bigtop-packages/src/deb/mahout/ bigtop-packages/src/rpm/mahout/SPECS/
Date Fri, 04 Nov 2011 00:18:02 GMT
Author: rvs
Date: Fri Nov  4 00:18:01 2011
New Revision: 1197368

URL: http://svn.apache.org/viewvc?rev=1197368&view=rev
Log:
Making Mahout compile

Modified:
    incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
    incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
    incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
    incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
    incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
    incubator/bigtop/branches/hadoop-0.23/bigtop.mk

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
(original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/do-component-build
Fri Nov  4 00:18:01 2011
@@ -16,5 +16,8 @@
 
 set -ex
 
-mvn clean install -Dmahout.skip.distribution=false -DskipTests -Dhadoop.version=0.22.0-SNAPSHOT
"$@"
-
+mvn clean install -Dmahout.skip.distribution=false -DskipTests -Dhadoop.version=0.23.0-SNAPSHOT
"$@"
+mkdir build
+for i in distribution/target/mahout*.tar.gz ; do
+  tar -C build --strip-components=1 -xzf $i
+done

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
(original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/install_mahout.sh
Fri Nov  4 00:18:01 2011
@@ -107,7 +107,7 @@ install -d -m 0755 $PREFIX/$DOC_DIR
 
 cp -ra ${BUILD_DIR}/lib/* $PREFIX/${LIB_DIR}/lib/
 cp ${BUILD_DIR}/mahout*.jar $PREFIX/$LIB_DIR
-cp ${BUILD_DIR}/mahout*.war $PREFIX/$LIB_DIR
+# cp ${BUILD_DIR}/mahout*.war $PREFIX/$LIB_DIR
 cp -a ${BUILD_DIR}/*.txt $PREFIX/$DOC_DIR
 cp -a ${BUILD_DIR}/bin/* $PREFIX/${LIB_DIR}/bin
 

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/mahout/patch Fri Nov
 4 00:18:01 2011
@@ -1,8 +1,357 @@
-Index: apache-mahout-c298f70/core/pom.xml
+Index: core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
 ===================================================================
---- apache-mahout-c298f70.orig/core/pom.xml	2011-09-29 15:22:25.369554983 -0700
-+++ apache-mahout-c298f70/core/pom.xml	2011-09-29 15:22:34.417110581 -0700
-@@ -140,10 +140,6 @@
+--- core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/Step1MapperTest.java
(working copy)
+@@ -22,6 +22,7 @@
+ import org.apache.hadoop.conf.Configuration;
+ import org.apache.hadoop.io.LongWritable;
+ import org.apache.hadoop.io.Text;
++import org.apache.hadoop.mapreduce.Mapper;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
+ import org.apache.mahout.common.MahoutTestCase;
+ import org.apache.mahout.common.RandomUtils;
+@@ -30,6 +31,7 @@
+ import org.apache.mahout.classifier.df.data.DataLoader;
+ import org.apache.mahout.classifier.df.data.Dataset;
+ import org.apache.mahout.classifier.df.data.Utils;
++import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
+ import org.apache.mahout.classifier.df.node.Leaf;
+ import org.apache.mahout.classifier.df.node.Node;
+ import org.junit.Test;
+@@ -109,8 +111,11 @@
+       // expected number of trees that this mapper will build
+       int mapNbTrees = Step1Mapper.nbTrees(NUM_MAPPERS, NUM_TREES, partition);
+ 
+-      MockContext context = new MockContext(new Step1Mapper(),
+-          new Configuration(), new TaskAttemptID(), mapNbTrees);
++      MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput> mockMapContextWrapper
=
++        new MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>();
++      Mapper<LongWritable,Text,TreeID,MapredOutput>.Context context =
++        mockMapContextWrapper.getMockContext(new Configuration(),
++            new TaskAttemptID(), mapNbTrees);
+ 
+       MockStep1Mapper mapper = new MockStep1Mapper(treeBuilder, dataset, seed,
+           partition, NUM_MAPPERS, NUM_TREES);
+@@ -127,10 +132,10 @@
+       mapper.cleanup(context);
+ 
+       // make sure the mapper built all its trees
+-      assertEquals(mapNbTrees, context.nbOutputs());
++      assertEquals(mapNbTrees, mockMapContextWrapper.nbOutputs());
+ 
+       // check the returned keys
+-      for (TreeID k : context.getKeys()) {
++      for (TreeID k : mockMapContextWrapper.getKeys()) {
+         assertEquals(partition, k.partition());
+         assertEquals(treeIndex, k.treeId());
+ 
+Index: core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
(revision 0)
++++ core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
(revision 0)
+@@ -0,0 +1,94 @@
++/**
++ * Licensed to the Apache Software Foundation (ASF) under one
++ * or more contributor license agreements.  See the NOTICE file
++ * distributed with this work for additional information
++ * regarding copyright ownership.  The ASF licenses this file
++ * to you under the Apache License, Version 2.0 (the
++ * "License"); you may not use this file except in compliance
++ * with the License.  You may obtain a copy of the License at
++ *
++ *     http://www.apache.org/licenses/LICENSE-2.0
++ *
++ * Unless required by applicable law or agreed to in writing, software
++ * distributed under the License is distributed on an "AS IS" BASIS,
++ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
++ * See the License for the specific language governing permissions and
++ * limitations under the License.
++ */
++
++package org.apache.mahout.classifier.df.mapreduce.partial;
++
++import java.io.IOException;
++
++import org.apache.hadoop.conf.Configuration;
++import org.apache.hadoop.mapreduce.Mapper;
++import org.apache.hadoop.mapreduce.TaskAttemptID;
++import org.apache.hadoop.mapreduce.lib.map.WrappedMapper;
++import org.apache.hadoop.mapreduce.task.MapContextImpl;
++import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
++
++public class MockMapContextWrapper<K1, V1, K2, V2> extends Mapper<K1, V1, K2, V2>
{
++  
++  private MockMapContext mockMapContext;
++  
++  public class MockMapContext extends MapContextImpl<K1, V1, K2, V2> {
++    
++    private final TreeID[] keys;
++    private final MapredOutput[] values;
++    private int index;
++
++    public MockMapContext(Configuration conf, TaskAttemptID taskAttemptID, int nbTrees)
{
++      super(conf, taskAttemptID, null, null, null, null, null);
++      
++      keys = new TreeID[nbTrees];
++      values = new MapredOutput[nbTrees];
++    }
++    
++    @Override
++    public void write(Object key, Object value) throws IOException {
++      if (index == keys.length) {
++        throw new IOException("Received more output than expected : " + index);
++      }
++
++      keys[index] = ((TreeID) key).clone();
++      values[index] = ((MapredOutput) value).clone();
++
++      index++;
++    }
++
++    /**
++     * @return number of outputs collected
++     */
++    public int nbOutputs() {
++      return index;
++    }
++
++    public TreeID[] getKeys() {
++      return keys;
++    }
++
++    public MapredOutput[] getValues() {
++      return values;
++    }
++  }
++
++  public Mapper<K1, V1, K2, V2>.Context getMockContext(Configuration conf, TaskAttemptID
taskAttemptID, int nbTrees) {
++    mockMapContext = new MockMapContext(conf, taskAttemptID, nbTrees);
++    return new WrappedMapper<K1, V1, K2, V2>().getMapContext(mockMapContext);
++  }
++
++  /**
++   * @return number of outputs collected
++   */
++  public int nbOutputs() {
++    return mockMapContext.nbOutputs();
++  }
++
++  public TreeID[] getKeys() {
++    return mockMapContext.getKeys();
++  }
++
++  public MapredOutput[] getValues() {
++    return mockMapContext.getValues();
++  }
++}
+
+Property changes on: core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockMapContextWrapper.java
+___________________________________________________________________
+Added: svn:eol-style
+   + native
+
+Index: core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/MockContext.java
(working copy)
+@@ -1,70 +0,0 @@
+-/**
+- * Licensed to the Apache Software Foundation (ASF) under one or more
+- * contributor license agreements.  See the NOTICE file distributed with
+- * this work for additional information regarding copyright ownership.
+- * The ASF licenses this file to You under the Apache License, Version 2.0
+- * (the "License"); you may not use this file except in compliance with
+- * the License.  You may obtain a copy of the License at
+- *
+- *     http://www.apache.org/licenses/LICENSE-2.0
+- *
+- * Unless required by applicable law or agreed to in writing, software
+- * distributed under the License is distributed on an "AS IS" BASIS,
+- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+- * See the License for the specific language governing permissions and
+- * limitations under the License.
+- */
+-package org.apache.mahout.classifier.df.mapreduce.partial;
+-
+-import java.io.IOException;
+-
+-import org.apache.hadoop.conf.Configuration;
+-import org.apache.hadoop.mapreduce.Mapper;
+-import org.apache.hadoop.mapreduce.TaskAttemptID;
+-import org.apache.hadoop.mapreduce.Mapper.Context;
+-import org.apache.mahout.classifier.df.mapreduce.MapredOutput;
+-
+-/**
+- * Special implementation that collects the output of the mappers
+- */
+-final class MockContext extends Context {
+-
+-  private final TreeID[] keys;
+-  private final MapredOutput[] values;
+-  private int index;
+-
+-  MockContext(Mapper<?,?,?,?> mapper, Configuration conf, TaskAttemptID taskid, int
nbTrees)
+-    throws IOException, InterruptedException {
+-    mapper.super(conf, taskid, null, null, null, null, null);
+-
+-    keys = new TreeID[nbTrees];
+-    values = new MapredOutput[nbTrees];
+-  }
+-
+-  @Override
+-  public void write(Object key, Object value) throws IOException {
+-    if (index == keys.length) {
+-      throw new IOException("Received more output than expected : " + index);
+-    }
+-
+-    keys[index] = ((TreeID) key).clone();
+-    values[index] = ((MapredOutput) value).clone();
+-
+-    index++;
+-  }
+-
+-  /**
+-   * @return number of outputs collected
+-   */
+-  public int nbOutputs() {
+-    return index;
+-  }
+-
+-  public TreeID[] getKeys() {
+-    return keys;
+-  }
+-
+-  public MapredOutput[] getValues() {
+-    return values;
+-  }
+-}
+Index: core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
(revision 1197192)
++++ core/src/test/java/org/apache/mahout/classifier/df/mapreduce/partial/PartialSequentialBuilder.java
(working copy)
+@@ -27,10 +27,12 @@
+ import org.apache.hadoop.io.Text;
+ import org.apache.hadoop.mapreduce.InputSplit;
+ import org.apache.hadoop.mapreduce.Job;
++import org.apache.hadoop.mapreduce.Mapper;
+ import org.apache.hadoop.mapreduce.RecordReader;
+ import org.apache.hadoop.mapreduce.TaskAttemptContext;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
+ import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
++import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl;
+ import org.apache.mahout.classifier.df.DFUtils;
+ import org.apache.mahout.classifier.df.DecisionForest;
+ import org.apache.mahout.classifier.df.builder.TreeBuilder;
+@@ -51,7 +53,7 @@
+ 
+   private static final Logger log = LoggerFactory.getLogger(PartialSequentialBuilder.class);
+ 
+-  private MockContext firstOutput;
++  private MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput> mockMapContextWrapper;
+ 
+   private final Dataset dataset;
+ 
+@@ -96,9 +98,11 @@
+ 
+     int numTrees = Builder.getNbTrees(conf); // total number of trees
+ 
+-    TaskAttemptContext task = new TaskAttemptContext(conf, new TaskAttemptID());
++    TaskAttemptContext task = new TaskAttemptContextImpl(conf, new TaskAttemptID());
+ 
+-    firstOutput = new MockContext(new Step1Mapper(), conf, task.getTaskAttemptID(), numTrees);
++    mockMapContextWrapper = new MockMapContextWrapper<LongWritable,Text,TreeID,MapredOutput>();
++    Mapper<LongWritable,Text,TreeID,MapredOutput>.Context firstOutput =
++      mockMapContextWrapper.getMockContext(conf, task.getTaskAttemptID(), numTrees);
+ 
+     /* first instance id in hadoop's order */
+     int[] firstIds = new int[nbSplits];
+@@ -143,7 +147,7 @@
+ 
+   @Override
+   protected DecisionForest parseOutput(Job job) throws IOException {
+-    return processOutput(firstOutput.getKeys(), firstOutput.getValues());
++    return processOutput(mockMapContextWrapper.getKeys(), mockMapContextWrapper.getValues());
+   }
+ 
+   /**
+Index: core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java	(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyStatusReporter.java	(working copy)
+@@ -55,4 +55,9 @@
+   public void setStatus(String status) {
+   }
+ 
++  @Override
++  public float getProgress() {
++    return 0;
++  }
++
+ }
+Index: core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java	(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyRecordWriter.java	(working copy)
+@@ -30,6 +30,10 @@
+ import org.apache.hadoop.mapreduce.Reducer;
+ import org.apache.hadoop.mapreduce.TaskAttemptContext;
+ import org.apache.hadoop.mapreduce.TaskAttemptID;
++import org.apache.hadoop.mapreduce.lib.map.WrappedMapper;
++import org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer;
++import org.apache.hadoop.mapreduce.task.MapContextImpl;
++import org.apache.hadoop.mapreduce.task.ReduceContextImpl;
+ 
+ public final class DummyRecordWriter<K, V> extends RecordWriter<K, V> {
+ 
+@@ -65,7 +69,8 @@
+                                                                       Configuration configuration,
+                                                                       RecordWriter<K2,
V2> output)
+     throws IOException, InterruptedException {
+-    return mapper.new Context(configuration, new TaskAttemptID(), null, output, null, new
DummyStatusReporter(), null);
++    return new WrappedMapper<K1, V1, K2, V2>().getMapContext(
++      new MapContextImpl<K1, V1, K2, V2>(configuration, new TaskAttemptID(), null,
output, null, new DummyStatusReporter(), null));
+   }
+ 
+   public static <K1, V1, K2, V2> Reducer<K1, V1, K2, V2>.Context build(Reducer<K1,
V1, K2, V2> reducer,
+@@ -74,7 +79,8 @@
+                                                                        Class<K1> keyClass,
+                                                                        Class<V1> valueClass)
+     throws IOException, InterruptedException {
+-    return reducer.new Context(configuration,
++    return new WrappedReducer<K1, V1, K2, V2>().getReducerContext(
++      new ReduceContextImpl<K1, V1, K2, V2>(configuration,
+                                new TaskAttemptID(),
+                                new MockIterator(),
+                                null,
+@@ -84,7 +90,7 @@
+                                new DummyStatusReporter(),
+                                null,
+                                keyClass,
+-                               valueClass);
++                               valueClass));
+   }
+ 
+ }
+Index: core/src/test/java/org/apache/mahout/common/DummyCounter.java
+===================================================================
+--- core/src/test/java/org/apache/mahout/common/DummyCounter.java	(revision 1197192)
++++ core/src/test/java/org/apache/mahout/common/DummyCounter.java	(working copy)
+@@ -19,8 +19,8 @@
+ 
+ package org.apache.mahout.common;
+ 
+-import org.apache.hadoop.mapreduce.Counter;
++import org.apache.hadoop.mapreduce.counters.GenericCounter;
+ 
+-final class DummyCounter extends Counter {
++final class DummyCounter extends GenericCounter {
+ 
+ }
+Index: core/pom.xml
+===================================================================
+--- core/pom.xml	(revision 1197192)
++++ core/pom.xml	(working copy)
+@@ -140,19 +140,6 @@
  
      <!-- Third Party -->
      <dependency>
@@ -10,10 +359,19 @@ Index: apache-mahout-c298f70/core/pom.xm
 -      <artifactId>hadoop-core</artifactId>
 -    </dependency>
 -    <dependency>
-       <groupId>org.codehaus.jackson</groupId>
-       <artifactId>jackson-core-asl</artifactId>
+-      <groupId>org.codehaus.jackson</groupId>
+-      <artifactId>jackson-core-asl</artifactId>
+-    </dependency>
+-    <dependency>
+-      <groupId>org.codehaus.jackson</groupId>
+-      <artifactId>jackson-mapper-asl</artifactId>
+-    </dependency>
+-
+-    <dependency>
+       <groupId>org.slf4j</groupId>
+       <artifactId>slf4j-api</artifactId>
      </dependency>
-@@ -211,4 +207,43 @@
+@@ -211,4 +198,51 @@
      </dependency>
  
    </dependencies>
@@ -47,25 +405,51 @@ Index: apache-mahout-c298f70/core/pom.xm
 +        </dependency>
 +        <dependency>
 +          <groupId>org.apache.hadoop</groupId>
-+          <artifactId>hadoop-hdfs</artifactId>
++          <artifactId>hadoop-mapreduce-client-common</artifactId>
 +        </dependency>
 +        <dependency>
 +          <groupId>org.apache.hadoop</groupId>
-+          <artifactId>hadoop-mapred</artifactId>
++          <artifactId>hadoop-mapreduce-client-core</artifactId>
++        </dependency>
++        <dependency>
++          <groupId>org.codehaus.jackson</groupId>
++          <artifactId>jackson-core-asl</artifactId>
++        </dependency>
++        <dependency>
++          <groupId>org.codehaus.jackson</groupId>
++          <artifactId>jackson-mapper-asl</artifactId>
 +        </dependency>
 +      </dependencies>
 +    </profile>
 +  </profiles>
  </project>
-Index: apache-mahout-c298f70/pom.xml
+Index: pom.xml
 ===================================================================
---- apache-mahout-c298f70.orig/pom.xml	2011-09-29 15:22:25.347055295 -0700
-+++ apache-mahout-c298f70/pom.xml	2011-09-29 15:22:29.555805208 -0700
-@@ -189,6 +189,220 @@
+--- pom.xml	(revision 1197192)
++++ pom.xml	(working copy)
+@@ -103,6 +103,17 @@
+     <url>https://issues.apache.org/jira/browse/MAHOUT</url>
+   </issueManagement>
+   
++  <repositories>
++    <repository>
++      <id>apache.snapshots</id>
++      <name>Apache Snapshot Repository</name>
++      <url>http://repository.apache.org/snapshots</url>
++      <releases>
++        <enabled>false</enabled>
++      </releases>
++    </repository>
++  </repositories>
++  
+   <dependencyManagement>
+     <dependencies>
+     
+@@ -260,6 +271,100 @@
+         </exclusions>
        </dependency>
- 
        <dependency>
-+         <groupId>org.apache.hadoop</groupId>
++        <groupId>org.apache.hadoop</groupId>
 +        <artifactId>hadoop-common</artifactId>
 +        <version>${hadoop.version}</version>
 +        <exclusions>
@@ -133,152 +517,32 @@ Index: apache-mahout-c298f70/pom.xml
 +            <groupId>org.eclipse.jdt</groupId>
 +            <artifactId>core</artifactId>
 +          </exclusion>
-+        </exclusions>
-+      </dependency>
-+      <dependency>
-+        <groupId>org.apache.hadoop</groupId>
-+        <artifactId>hadoop-hdfs</artifactId>
-+        <version>${hadoop.version}</version>
-+        <exclusions>
-+          <exclusion>
-+            <groupId>net.sf.kosmosfs</groupId>
-+            <artifactId>kfs</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jetty</artifactId>
-+          </exclusion>
 +          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jetty-util</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>hsqldb</groupId>
-+            <artifactId>hsqldb</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>commons-el</groupId>
-+            <artifactId>commons-el</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>junit</groupId>
-+            <artifactId>junit</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>oro</groupId>
-+            <artifactId>oro</artifactId>
++            <groupId>org.slf4j</groupId>
++            <artifactId>slf4j-api</artifactId>
 +          </exclusion>
 +          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jsp-2.1</artifactId>
++            <groupId>org.slf4j</groupId>
++            <artifactId>slf4j-jcl</artifactId>
 +          </exclusion>
 +          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jsp-api-2.1</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>servlet-api-2.5</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>commons-net</groupId>
-+            <artifactId>commons-net</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>tomcat</groupId>
-+            <artifactId>jasper-runtime</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>tomcat</groupId>
-+            <artifactId>jasper-compiler</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>xmlenc</groupId>
-+            <artifactId>xmlenc</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>net.java.dev.jets3t</groupId>
-+            <artifactId>jets3t</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.eclipse.jdt</groupId>
-+            <artifactId>core</artifactId>
++            <groupId>org.slf4j</groupId>
++            <artifactId>slf4j-log4j12</artifactId>
 +          </exclusion>
 +        </exclusions>
 +      </dependency>
 +      <dependency>
 +        <groupId>org.apache.hadoop</groupId>
-+        <artifactId>hadoop-mapred</artifactId>
++        <artifactId>hadoop-mapreduce-client-core</artifactId>
++        <version>${hadoop.version}</version>
++      </dependency>
++      <dependency>
++        <groupId>org.apache.hadoop</groupId>
++        <artifactId>hadoop-mapreduce-client-common</artifactId>
 +        <version>${hadoop.version}</version>
-+        <exclusions>
-+          <exclusion>
-+            <groupId>net.sf.kosmosfs</groupId>
-+            <artifactId>kfs</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jetty</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jetty-util</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>hsqldb</groupId>
-+            <artifactId>hsqldb</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>commons-el</groupId>
-+            <artifactId>commons-el</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>junit</groupId>
-+            <artifactId>junit</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>oro</groupId>
-+            <artifactId>oro</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jsp-2.1</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>jsp-api-2.1</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.mortbay.jetty</groupId>
-+            <artifactId>servlet-api-2.5</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>commons-net</groupId>
-+            <artifactId>commons-net</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>tomcat</groupId>
-+            <artifactId>jasper-runtime</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>tomcat</groupId>
-+            <artifactId>jasper-compiler</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>xmlenc</groupId>
-+            <artifactId>xmlenc</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>net.java.dev.jets3t</groupId>
-+            <artifactId>jets3t</artifactId>
-+          </exclusion>
-+          <exclusion>
-+            <groupId>org.eclipse.jdt</groupId>
-+            <artifactId>core</artifactId>
-+          </exclusion>
-+        </exclusions>
 +      </dependency>
 +
 +      <dependency>
-         <groupId>org.apache.hadoop</groupId>
-         <artifactId>hadoop-core</artifactId>
-         <version>0.20.204.0</version>
+         <groupId>org.codehaus.jackson</groupId>
+         <artifactId>jackson-core-asl</artifactId>
+         <version>1.8.2</version>

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/deb/mahout/rules Fri Nov  4
00:18:01 2011
@@ -25,7 +25,7 @@ export DH_OPTIONS
 
 patch: patch-stamp
 patch-stamp:	
-	patch -p1 < debian/patch
+	patch -p0 < debian/patch
 	touch $@
 
 clean:	
@@ -40,9 +40,6 @@ build-indep: build-indep-stamp
 build-indep-stamp: patch-stamp
 	# we'll just use the build from the tarball.
 	bash debian/do-component-build -Dmaven.repo.local=${HOME}/.m2/repository
-	mkdir -p debian/tmp
-	(cd distribution/target/mahout-distribution-${MAHOUT_BASE_VERSION}/mahout-distribution-${MAHOUT_BASE_VERSION}
&& tar cf - --exclude=debian/\* .) | (cd debian/tmp && tar xf -)
-#	(cd debian/tmp && tar xvzf ../../distribution/target/mahout-distribution-${MAHOUT_BASE_VERSION}.tar.gz)
 	touch $@
 
 install: install-indep
@@ -50,7 +47,7 @@ install-indep:	
 	dh_testdir
 	dh_testroot
 	sh -x debian/install_mahout.sh \
-	  --build-dir=debian/tmp \
+	  --build-dir=build \
           --doc-dir=/usr/share/doc/mahout \
 	  --prefix=debian/mahout
 	rm debian/mahout/usr/lib/mahout/lib/hadoop-*.jar

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
(original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
Fri Nov  4 00:18:01 2011
@@ -66,8 +66,8 @@ diverse community to facilitate discussi
 also on potential use cases. Come to the mailing lists to find out more.
     
 %prep
-%setup -n apache-mahout-c298f70
-%patch0 -p1 
+%setup -n apache-mahout-464be41
+%patch0 -p0
 
 %build
 bash $RPM_SOURCE_DIR/do-component-build
@@ -75,7 +75,7 @@ bash $RPM_SOURCE_DIR/do-component-build
 %install
 %__rm -rf $RPM_BUILD_ROOT
 sh $RPM_SOURCE_DIR/install_mahout.sh \
-          --build-dir=distribution/target/mahout-distribution-%{mahout_base_version}/mahout-distribution-%{mahout_base_version}
\
+          --build-dir=build \
           --prefix=$RPM_BUILD_ROOT \
           --doc-dir=%{doc_mahout} 
 rm -f $RPM_BUILD_ROOT/usr/lib/mahout/lib/hadoop*.jar

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop.mk
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop.mk?rev=1197368&r1=1197367&r2=1197368&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop.mk (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop.mk Fri Nov  4 00:18:01 2011
@@ -128,7 +128,7 @@ MAHOUT_RELEASE_VERSION=1
 MAHOUT_TARBALL_DST=mahout-distribution-$(MAHOUT_BASE_VERSION)-src.tar.gz
 #MAHOUT_TARBALL_SRC=$(MAHOUT_TARBALL_DST)
 #MAHOUT_SITE=$(APACHE_MIRROR)/mahout/0.5/
-MAHOUT_TARBALL_SRC=c298f70
+MAHOUT_TARBALL_SRC=464be41
 MAHOUT_SITE=https://github.com/apache/mahout/tarball
 $(eval $(call PACKAGE,mahout,MAHOUT))
 



Mime
View raw message