incubator-bigtop-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From r..@apache.org
Subject svn commit: r1242013 - in /incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources: package_data_apt.xml package_data_urpmi.xml package_data_yum.xml package_data_zypper.xml
Date Wed, 08 Feb 2012 17:54:19 GMT
Author: rvs
Date: Wed Feb  8 17:54:19 2012
New Revision: 1242013

URL: http://svn.apache.org/viewvc?rev=1242013&view=rev
Log:
BIGTOP-391. update package test manifest

Modified:
    incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
    incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
    incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
    incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml

Modified: incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml?rev=1242013&r1=1242012&r2=1242013&view=diff
==============================================================================
--- incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
(original)
+++ incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_apt.xml
Wed Feb  8 17:54:19 2012
@@ -65,7 +65,7 @@
   * Smart defaults for services. You can get a properly configured system
     running quickly, while still being able to override settings as needed.
     </description>
-    <url>http://incubator.apache.org/whirr</url>
+    <url>http://whirr.apache.org/</url>
   </metadata>
   <deps>
     <sun-java6-jre/>
@@ -1139,36 +1139,6 @@
     </flume-conf>
   </alternatives>
 </flume>
-<flume-master>
-  <metadata>
-    <summary>central administration point for the flume data collection system</summary>
-    <description>The Flume master daemon is the central administration and data path
control
- point for flume nodes.</description>
-    <url>http://www.cloudera.com</url>
-  </metadata>
-  <deps>
-    <flume>/self</flume>
-  </deps>
-  <services>
-    <flume-master>
-       <runlevel>2</runlevel><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
-       <oninstall>start</oninstall>
-       <configured>true</configured>
-    </flume-master>
-  </services>
-  <content>
-    <file name="/." owners="0" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/etc" owners="119" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/etc/init.d" owners="32" perm="drwxr-xr-x" user="root" group="root"/>
-    <config name="/etc/init.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root" group="root"/>
-    <file name="/usr" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/share" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/share/doc" owners="418" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/share/doc/flume-master" owners="1" perm="drwxr-xr-x" user="root"
group="root"/>
-    <file name="/usr/share/doc/flume-master/copyright" owners="1" perm="-rw-r--r--" user="root"
group="root"/>
-    <file name="/usr/share/doc/flume-master/changelog.Debian.gz" owners="1" perm="-rw-r--r--"
user="root" group="root"/>
-  </content>
-</flume-master>
 <flume-node>
   <metadata>
     <summary>core element of Flume's data path that collects and delivers data</summary>
@@ -1446,7 +1416,7 @@
  In case of workflow job failure, the workflow job can be rerun skipping
  previously completed actions, the workflow application can be patched before
  being rerun.</description>
-    <url>http://archive.cloudera.com/cdh/3/oozie</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <deps>
     <oozie-client>/self</oozie-client>
@@ -1575,7 +1545,7 @@
  user can deploy workflows and perform other administrative and
  monitoring tasks such as start, stop, kill, resume workflows
  and coordinator jobs.</description>
-    <url>http://archive.cloudera.com/cdh/3/oozie</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <content>
     <file name="/." owners="0" perm="drwxr-xr-x" user="root" group="root"/>
@@ -2390,7 +2360,7 @@
     <summary>A high-performance coordination service for distributed applications.</summary>
     <description>ZooKeeper is a centralized service for maintaining configuration information,
naming, providing distributed synchronization, and providing group services.  All of these
kinds of services are used in some form or another by distributed applications. Each time
they are implemented there is a lot of work that goes into fixing the bugs and race conditions
that are inevitable. Because of the difficulty of implementing these kinds of services, applications
initially usually skimp on them ,which make them brittle in the presence of change and difficult
to manage. Even when done correctly, different implementations of these services lead to management
complexity when the applications are deployed.
     </description>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -2787,7 +2757,7 @@
     <!-- license>APL2</license -->
     <!-- arch>universal</arch -->
     <summary>This runs the zookeeper server on startup.</summary>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- vendor>(none)</vendor -->
     <!-- group>Development/Libraries</group -->
     <!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends
-->
@@ -2842,7 +2812,7 @@
     automatically, allowing the user to focus on semantics rather than efficiency.
  * Extensibility
     Users can create their own functions to do special-purpose processing.</description>
-    <url>http://hadoop.apache.org/pig/</url>
+    <url>http://pig.apache.org/</url>
   </metadata>
   <deps>
     <sun-java6-jre/>
@@ -6787,7 +6757,7 @@
  plug in their custom mappers and reducers to do more sophisticated
  analysis which may not be supported by the built-in capabilities of
  the language.</description>
-    <url>http://hadoop.apache.org/hive/</url>
+    <url>http://hive.apache.org/</url>
   </metadata>
   <deps>
     <sun-java6-jre/>
@@ -7099,7 +7069,7 @@
     <description>Use it when you need random, realtime read/write access to your Big
Data.
  This project's goal is the hosting of very large tables -- billions of rows
  X millions of columns -- atop clusters of commodity hardware.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -7290,7 +7260,7 @@
   <metadata>
     <summary>Documentation for HBase</summary>
     <description>This package contains the HBase manual and JavaDoc.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <content>
@@ -9065,7 +9035,7 @@
   <metadata>
     <summary>HMaster is the "master server" for a HBase</summary>
     <description>There is only one HMaster for a single HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>
@@ -9098,7 +9068,7 @@
     <summary>HRegionServer makes a set of HRegions available to clients</summary>
     <description>It checks in with the HMaster. There are many HRegionServers in a
single
  HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>
@@ -9131,7 +9101,7 @@
     <summary>Provides an HBase Thrift service</summary>
     <description>This package provides a Thrift service interface to the HBase distributed
  database.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>

Modified: incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml?rev=1242013&r1=1242012&r2=1242013&view=diff
==============================================================================
--- incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
(original)
+++ incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_urpmi.xml
Wed Feb  8 17:54:19 2012
@@ -63,7 +63,7 @@ also on potential use cases. Come to the
   service.
 * Smart defaults for services. You can get a properly configured system
   running quickly, while still being able to override settings as needed.</description>
-    <url>http://incubator.apache.org/whirr</url>
+    <url>http://whirr.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -804,7 +804,7 @@ also on potential use cases. Come to the
  mechanisms and many failover and recovery mechanisms.  The system is centrally
  managed and allows for intelligent dynamic management.  It uses a simple
  extensible data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
+    <url>http://incubator.apache.org/projects/flume.html</url>
   </metadata>
   <deps>
     <tag name="/bin/sh"/>
@@ -1117,44 +1117,11 @@ also on potential use cases. Come to the
     </flume-conf>
   </alternatives>
 </flume>
-<flume-master>
-  <metadata>
-    <summary>The flume master daemon is the central administration and data path control
point for flume nodes.</summary>
-    <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
-  </metadata>
-  <deps>
-    <tag name="/bin/bash"/>
-    <tag name="/bin/sh"/>
-    <tag name="/sbin/chkconfig"/>
-    <tag name="/sbin/service"/>
-    <tag name="/usr/bin/env"/>
-    <tag name="/usr/sbin/useradd"/>
-    <flume>/self</flume>
-    <jre>>=1.6</jre>
-    <redhat-lsb/>
-    <sh-utils/>
-    <textutils/>
-  </deps>
-  <services>
-    <flume-master>
-       <runlevel>2</runlevel><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
-       <oninstall>stop</oninstall>
-       <configured>true</configured>
-    </flume-master>
-  </services>
-  <content>
-    <file name="/etc/rc.d/init.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root"
group="root"/>
-    <file name="/usr/lib/flume" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/lib/flume/bin" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/lib/flume/bin/flume-daemon.sh" owners="2" perm="-rwxr-xr-x" user="root"
group="root"/>
-  </content>
-</flume-master>
 <flume-node>
   <metadata>
     <summary>The flume node daemon is a core element of flume's data path and is responsible
for generating, processing, and delivering data.</summary>
     <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
+    <url>http://incubator.apache.org/projects/flume.html</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -1418,7 +1385,7 @@ also on potential use cases. Come to the
  In case of workflow job failure, the workflow job can be rerun skipping
  previously completed actions, the workflow application can be patched before
  being rerun.</description>
-    <url>http://www.cloudera.com</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -1542,7 +1509,7 @@ also on potential use cases. Come to the
  you can also change the status of the entire system, get vesion
  information. This client utility also allows you to validate
  any worflows before they are deployed to the Oozie server.</description>
-    <url>http://www.cloudera.com</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -2356,7 +2323,7 @@ into fixing the bugs and race conditions
 difficulty of implementing these kinds of services, applications initially
 usually skimp on them ,which make them brittle in the presence of change and
 difficult to manage. Even when done correctly, different implementations of these services
lead to management complexity when the applications are deployed.</description>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -2740,7 +2707,7 @@ difficult to manage. Even when done corr
     <!-- arch>universal</arch -->
     <summary>The Hadoop Zookeeper server</summary>
     <description>This package starts the zookeeper server on startup</description>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- vendor>(none)</vendor -->
     <!-- group>Development/Libraries</group -->
     <!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends
-->
@@ -2788,7 +2755,7 @@ difficult to manage. Even when done corr
    automatically, allowing the user to focus on semantics rather than efficiency.
  * Extensibility
    Users can create their own functions to do special-purpose processing.</description>
-    <url>http://hadoop.apache.org/pig/</url>
+    <url>http://pig.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/sh"/>
@@ -6717,7 +6684,7 @@ difficult to manage. Even when done corr
  plug in their custom mappers and reducers to do more sophisticated
  analysis which may not be supported by the built-in capabilities of
  the language.</description>
-    <url>http://hadoop.apache.org/hive/</url>
+    <url>http://hive.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/sh"/>
@@ -7157,7 +7124,7 @@ difficult to manage. Even when done corr
     * Cascading source and sink modules
     * Extensible jruby-based (JIRB) shell
     * Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia;
or via JMX</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -7336,7 +7303,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>Hbase Documentation</summary>
     <description>Documentation for Hbase</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <content>
@@ -9105,7 +9072,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase master Server.</summary>
     <description>HMaster is the "master server" for a HBase. There is only one HMaster
for a single HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -9128,7 +9095,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase RegionServer server.</summary>
     <description>HRegionServer makes a set of HRegions available to clients. It checks
in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -9151,7 +9118,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase Thrift Interface</summary>
     <description>ThriftServer - this class starts up a Thrift server which implements
the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for
scalable cross-language services development. It combines a powerful software stack with a
code generation engine to build services that work efficiently and seamlessly between C++,
Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it
as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook
has announced their intent to migrate Thrift into Apache Incubator.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>

Modified: incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml?rev=1242013&r1=1242012&r2=1242013&view=diff
==============================================================================
--- incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
(original)
+++ incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_yum.xml
Wed Feb  8 17:54:19 2012
@@ -156,7 +156,7 @@ also on potential use cases. Come to the
   service.
 * Smart defaults for services. You can get a properly configured system
   running quickly, while still being able to override settings as needed.</description>
-      <url>http://incubator.apache.org/whirr</url>
+      <url>http://whirr.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -1060,7 +1060,7 @@ also on potential use cases. Come to the
  mechanisms and many failover and recovery mechanisms.  The system is centrally
  managed and allows for intelligent dynamic management.  It uses a simple
  extensible data model that allows for online analytic applications.</description>
-      <url>https://github.com/cloudera/flume</url>
+      <url>http://incubator.apache.org/projects/flume.html</url>
     </metadata>
     <deps>
       <tag name="/bin/sh"/>
@@ -1369,46 +1369,11 @@ also on potential use cases. Come to the
       </flume-conf>
     </alternatives>
   </flume>
-  <flume-master>
-    <metadata>
-      <summary>The flume master daemon is the central administration and data path
control point for flume nodes.</summary>
-      <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-      <url>https://github.com/cloudera/flume</url>
-    </metadata>
-    <deps>
-      <tag name="/bin/bash"/>
-      <tag name="/bin/sh"/>
-      <tag name="/sbin/chkconfig"/>
-      <tag name="/sbin/service"/>
-      <tag name="/usr/bin/env"/>
-      <tag name="/usr/sbin/useradd"/>
-      <flume>/self</flume>
-      <jre>&gt;=1.6</jre>
-      <redhat-lsb/>
-      <sh-utils/>
-      <textutils/>
-    </deps>
-    <services>
-      <flume-master>
-        <runlevel>2</runlevel>
-        <runlevel>3</runlevel>
-        <runlevel>4</runlevel>
-        <runlevel>5</runlevel>
-        <oninstall>stop</oninstall>
-        <configured>true</configured>
-      </flume-master>
-    </services>
-    <content>
-      <file name="/etc/rc.d/init.d/flume-master" perm="-rwxr-xr-x" group="root" owners="-1"
user="root"/>
-      <file name="/usr/lib/flume" perm="drwxr-xr-x" group="root" owners="-1" user="root"/>
-      <file name="/usr/lib/flume/bin" perm="drwxr-xr-x" group="root" owners="-1" user="root"/>
-    </content>
-  </flume-master>
   <flume-node>
     <metadata>
       <summary>The flume node daemon is a core element of flume's data path and is
responsible for generating, processing, and delivering data.</summary>
       <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-      <url>https://github.com/cloudera/flume</url>
+      <url>http://incubator.apache.org/projects/flume.html</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -1685,7 +1650,7 @@ also on potential use cases. Come to the
  In case of workflow job failure, the workflow job can be rerun skipping
  previously completed actions, the workflow application can be patched before
  being rerun.</description>
-      <url>http://www.cloudera.com</url>
+      <url>http://incubator.apache.org/oozie/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -1815,7 +1780,7 @@ also on potential use cases. Come to the
  you can also change the status of the entire system, get vesion
  information. This client utility also allows you to validate
  any worflows before they are deployed to the Oozie server.</description>
-      <url>http://www.cloudera.com</url>
+      <url>http://incubator.apache.org/oozie/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -4171,7 +4136,7 @@ into fixing the bugs and race conditions
 difficulty of implementing these kinds of services, applications initially
 usually skimp on them ,which make them brittle in the presence of change and
 difficult to manage. Even when done correctly, different implementations of these services
lead to management complexity when the applications are deployed.</description>
-      <url>http://hadoop.apache.org/zookeeper/</url>
+      <url>http://zookeeper.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/sh"/>
@@ -4550,7 +4515,7 @@ difficult to manage. Even when done corr
     <metadata>
       <summary>The Hadoop Zookeeper server</summary>
       <description>This package starts the zookeeper server on startup</description>
-      <url>http://hadoop.apache.org/zookeeper/</url>
+      <url>http://zookeeper.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -4595,7 +4560,7 @@ difficult to manage. Even when done corr
    automatically, allowing the user to focus on semantics rather than efficiency.
  * Extensibility
    Users can create their own functions to do special-purpose processing.</description>
-      <url>http://hadoop.apache.org/pig/</url>
+      <url>http://pig.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/sh"/>
@@ -8471,7 +8436,7 @@ difficult to manage. Even when done corr
  plug in their custom mappers and reducers to do more sophisticated
  analysis which may not be supported by the built-in capabilities of
  the language.</description>
-      <url>http://hadoop.apache.org/hive/</url>
+      <url>http://hive.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/sh"/>
@@ -8927,7 +8892,7 @@ difficult to manage. Even when done corr
     * Cascading source and sink modules
     * Extensible jruby-based (JIRB) shell
     * Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia;
or via JMX</description>
-      <url>http://hadoop.apache.org/hbase/</url>
+      <url>http://hbase.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/sh"/>
@@ -9114,7 +9079,7 @@ difficult to manage. Even when done corr
     <metadata>
       <summary>Hbase Documentation</summary>
       <description>Documentation for Hbase</description>
-      <url>http://hadoop.apache.org/hbase/</url>
+      <url>http://hbase.apache.org/</url>
     </metadata>
     <content>
       <file name="/usr/share/doc/hbase-0.90.4.16" perm="drwxr-xr-x" group="root" owners="-1"
user="root"/>
@@ -11842,7 +11807,7 @@ difficult to manage. Even when done corr
     <metadata>
       <summary>The Hadoop HBase master Server.</summary>
       <description>HMaster is the "master server" for a HBase. There is only one HMaster
for a single HBase deployment.</description>
-      <url>http://hadoop.apache.org/hbase/</url>
+      <url>http://hbase.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -11868,7 +11833,7 @@ difficult to manage. Even when done corr
     <metadata>
       <summary>The Hadoop HBase RegionServer server.</summary>
       <description>HRegionServer makes a set of HRegions available to clients. It checks
in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
-      <url>http://hadoop.apache.org/hbase/</url>
+      <url>http://hbase.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>
@@ -11894,7 +11859,7 @@ difficult to manage. Even when done corr
     <metadata>
       <summary>The Hadoop HBase Thrift Interface</summary>
       <description>ThriftServer - this class starts up a Thrift server which implements
the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for
scalable cross-language services development. It combines a powerful software stack with a
code generation engine to build services that work efficiently and seamlessly between C++,
Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it
as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook
has announced their intent to migrate Thrift into Apache Incubator.</description>
-      <url>http://hadoop.apache.org/hbase/</url>
+      <url>http://hbase.apache.org/</url>
     </metadata>
     <deps>
       <tag name="/bin/bash"/>

Modified: incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml
URL: http://svn.apache.org/viewvc/incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml?rev=1242013&r1=1242012&r2=1242013&view=diff
==============================================================================
--- incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml
(original)
+++ incubator/bigtop/trunk/bigtop-tests/test-artifacts/package/src/main/resources/package_data_zypper.xml
Wed Feb  8 17:54:19 2012
@@ -63,7 +63,7 @@ also on potential use cases. Come to the
   service.
 * Smart defaults for services. You can get a properly configured system
   running quickly, while still being able to override settings as needed.</description>
-    <url>http://incubator.apache.org/whirr</url>
+    <url>http://whirr.apache.org/</url>
   </metadata>
   <deps>
     <jre>>=1.6</jre>
@@ -804,7 +804,7 @@ also on potential use cases. Come to the
  mechanisms and many failover and recovery mechanisms.  The system is centrally
  managed and allows for intelligent dynamic management.  It uses a simple
  extensible data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
+    <url>http://incubator.apache.org/projects/flume.html</url>
   </metadata>
   <deps>
     <sh-utils/>
@@ -1117,44 +1117,11 @@ also on potential use cases. Come to the
     </flume-conf>
   </alternatives>
 </flume>
-<flume-master>
-  <metadata>
-    <summary>The flume master daemon is the central administration and data path control
point for flume nodes.</summary>
-    <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
-  </metadata>
-  <deps>
-    <flume>/self</flume>
-    <sh-utils/>
-    <textutils/>
-    <tag name="/usr/sbin/useradd"/>
-    <tag name="/sbin/chkconfig"/>
-    <tag name="/sbin/service"/>
-    <jre>>=1.6</jre>
-    <insserv/>
-    <tag name="/bin/sh"/>
-    <tag name="/bin/bash"/>
-    <tag name="/usr/bin/env"/>
-  </deps>
-  <services>
-    <flume-master>
-       <!-- BUG https://jira.cloudera.com/browse/KITCHEN-1095 --><runlevel>3</runlevel><runlevel>4</runlevel><runlevel>5</runlevel>
-       <oninstall>stop</oninstall>
-       <configured>true</configured>
-    </flume-master>
-  </services>
-  <content>
-    <file name="/etc/rc.d/flume-master" owners="1" perm="-rwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/lib/flume" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/lib/flume/bin" owners="2" perm="drwxr-xr-x" user="root" group="root"/>
-    <file name="/usr/lib/flume/bin/flume-daemon.sh" owners="2" perm="-rwxr-xr-x" user="root"
group="root"/>
-  </content>
-</flume-master>
 <flume-node>
   <metadata>
     <summary>The flume node daemon is a core element of flume's data path and is responsible
for generating, processing, and delivering data.</summary>
     <description>Flume is a reliable, scalable, and manageable distributed data collection
application for collecting data such as logs and delivering it to data stores such as Hadoop's
HDFS. It can efficiently collect, aggregate, and move large amounts of log data. It has a
simple, but flexible, architecture based on streaming data flows. It is robust and fault tolerant
with tunable reliability mechanisms and many failover and recovery mechanisms. The system
is centrally managed and allows for intelligent dynamic management. It uses a simple extensible
data model that allows for online analytic applications.</description>
-    <url>https://github.com/cloudera/flume</url>
+    <url>http://incubator.apache.org/projects/flume.html</url>
   </metadata>
   <deps>
     <flume>/self</flume>
@@ -1418,7 +1385,7 @@ also on potential use cases. Come to the
  In case of workflow job failure, the workflow job can be rerun skipping
  previously completed actions, the workflow application can be patched before
  being rerun.</description>
-    <url>http://www.cloudera.com</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <deps>
     <tag name="/usr/sbin/groupadd"/>
@@ -1543,7 +1510,7 @@ also on potential use cases. Come to the
  you can also change the status of the entire system, get vesion
  information. This client utility also allows you to validate
  any worflows before they are deployed to the Oozie server.</description>
-    <url>http://www.cloudera.com</url>
+    <url>http://incubator.apache.org/oozie/</url>
   </metadata>
   <deps>
     <tag name="/bin/bash"/>
@@ -2356,7 +2323,7 @@ into fixing the bugs and race conditions
 difficulty of implementing these kinds of services, applications initially
 usually skimp on them ,which make them brittle in the presence of change and
 difficult to manage. Even when done correctly, different implementations of these services
lead to management complexity when the applications are deployed.</description>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -2728,9 +2695,9 @@ difficult to manage. Even when done corr
   <alternatives>
     <zookeeper-conf>
       <status>auto</status>
-      <link>/etc/zookeeper</link>
-      <value>/etc/zookeeper.dist</value>
-      <alt>/etc/zookeeper.dist</alt>
+      <link>/etc/zookeeper/conf</link>
+      <value>/etc/zookeeper/conf.dist</value>
+      <alt>/etc/zookeeper/conf.dist</alt>
     </zookeeper-conf>
   </alternatives>
 </zookeeper>
@@ -2740,7 +2707,7 @@ difficult to manage. Even when done corr
     <!-- arch>universal</arch -->
     <summary>The Hadoop Zookeeper server</summary>
     <description>This package starts the zookeeper server on startup</description>
-    <url>http://hadoop.apache.org/zookeeper/</url>
+    <url>http://zookeeper.apache.org/</url>
     <!-- vendor>(none)</vendor -->
     <!-- group>Development/Libraries</group -->
     <!-- depends><dep>adduser</dep><dep>sun-java6-jre</dep><dep>sun-java6-bin</dep></depends
-->
@@ -2788,7 +2755,7 @@ difficult to manage. Even when done corr
    automatically, allowing the user to focus on semantics rather than efficiency.
  * Extensibility
    Users can create their own functions to do special-purpose processing.</description>
-    <url>http://hadoop.apache.org/pig/</url>
+    <url>http://pig.apache.org/</url>
   </metadata>
   <deps>
     <hadoop/>
@@ -6696,12 +6663,12 @@ difficult to manage. Even when done corr
     <doc  name="/usr/share/man/man1/pig.1.gz" owners="1" perm="-rw-r--r--" user="root"
group="root"/>
   </content>
   <alternatives>
-    <pig> <!-- BUG: https://issues.cloudera.org/browse/DISTRO-223 -->
+    <pig-conf>
       <status>auto</status>
       <link>/etc/pig/conf</link>
       <value>/etc/pig/conf.dist</value>
       <alt>/etc/pig/conf.dist</alt>
-    </pig>
+    </pig-conf>
   </alternatives>
 </pig>
 <hive>
@@ -6717,7 +6684,7 @@ difficult to manage. Even when done corr
  plug in their custom mappers and reducers to do more sophisticated
  analysis which may not be supported by the built-in capabilities of
  the language.</description>
-    <url>http://hadoop.apache.org/hive/</url>
+    <url>http://hive.apache.org/</url>
   </metadata>
   <deps>
     <hadoop>>=0.20.1</hadoop>
@@ -7077,7 +7044,7 @@ difficult to manage. Even when done corr
     * Cascading source and sink modules
     * Extensible jruby-based (JIRB) shell
     * Support for exporting metrics via the Hadoop metrics subsystem to files or Ganglia;
or via JMX</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <deps>
@@ -7256,7 +7223,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>Hbase Documentation</summary>
     <description>Documentation for Hbase</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
     <!-- group>misc</group -->
   </metadata>
   <content>
@@ -9025,7 +8992,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase master Server.</summary>
     <description>HMaster is the "master server" for a HBase. There is only one HMaster
for a single HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>
@@ -9048,7 +9015,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase RegionServer server.</summary>
     <description>HRegionServer makes a set of HRegions available to clients. It checks
in with the HMaster. There are many HRegionServers in a single HBase deployment.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>
@@ -9071,7 +9038,7 @@ difficult to manage. Even when done corr
   <metadata>
     <summary>The Hadoop HBase Thrift Interface</summary>
     <description>ThriftServer - this class starts up a Thrift server which implements
the Hbase API specified in the Hbase.thrift IDL file. "Thrift is a software framework for
scalable cross-language services development. It combines a powerful software stack with a
code generation engine to build services that work efficiently and seamlessly between C++,
Java, Python, PHP, and Ruby. Thrift was developed at Facebook, and we are now releasing it
as open source." For additional information, see http://developers.facebook.com/thrift/. Facebook
has announced their intent to migrate Thrift into Apache Incubator.</description>
-    <url>http://hadoop.apache.org/hbase/</url>
+    <url>http://hbase.apache.org/</url>
   </metadata>
   <deps>
     <hbase>/self</hbase>



Mime
View raw message