falcon-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From srik...@apache.org
Subject [15/47] Fixes for Checkstyle
Date Fri, 26 Apr 2013 15:50:31 GMT
http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/entity/store/ConfigurationStoreTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/entity/store/ConfigurationStoreTest.java b/common/src/test/java/org/apache/falcon/entity/store/ConfigurationStoreTest.java
index 15e6e32..e58184e 100644
--- a/common/src/test/java/org/apache/falcon/entity/store/ConfigurationStoreTest.java
+++ b/common/src/test/java/org/apache/falcon/entity/store/ConfigurationStoreTest.java
@@ -18,12 +18,12 @@
 
 package org.apache.falcon.entity.store;
 
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
 import org.apache.falcon.entity.v0.EntityType;
 import org.apache.falcon.entity.v0.process.Process;
 import org.apache.falcon.util.StartupProperties;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 import org.testng.Assert;
 import org.testng.annotations.AfterSuite;
@@ -34,49 +34,49 @@ import java.io.IOException;
 
 public class ConfigurationStoreTest {
 
-  private static Logger LOG = Logger.getLogger(ConfigurationStoreTest.class);
+    private static Logger LOG = Logger.getLogger(ConfigurationStoreTest.class);
 
-  private ConfigurationStore store = ConfigurationStore.get();
+    private ConfigurationStore store = ConfigurationStore.get();
 
-  @Test
-  public void testPublish() throws Exception {
-    Process process = new Process();
-    process.setName("hello");
-    store.publish(EntityType.PROCESS, process);
-    Process p = store.get(EntityType.PROCESS, "hello");
-    Assert.assertEquals(p, process);
-  }
+    @Test
+    public void testPublish() throws Exception {
+        Process process = new Process();
+        process.setName("hello");
+        store.publish(EntityType.PROCESS, process);
+        Process p = store.get(EntityType.PROCESS, "hello");
+        Assert.assertEquals(p, process);
+    }
 
-  @Test
-  public void testGet() throws Exception {
-    Process p = store.get(EntityType.PROCESS, "notfound");
-    Assert.assertNull(p);
-  }
+    @Test
+    public void testGet() throws Exception {
+        Process p = store.get(EntityType.PROCESS, "notfound");
+        Assert.assertNull(p);
+    }
 
-  @Test
-  public void testRemove() throws Exception {
-    Process process = new Process();
-    process.setName("remove");
-    store.publish(EntityType.PROCESS, process);
-    Process p = store.get(EntityType.PROCESS, "remove");
-    Assert.assertEquals(p, process);
-    store.remove(EntityType.PROCESS, "remove");
-    p = store.get(EntityType.PROCESS, "remove");
-    Assert.assertNull(p);
-  }
+    @Test
+    public void testRemove() throws Exception {
+        Process process = new Process();
+        process.setName("remove");
+        store.publish(EntityType.PROCESS, process);
+        Process p = store.get(EntityType.PROCESS, "remove");
+        Assert.assertEquals(p, process);
+        store.remove(EntityType.PROCESS, "remove");
+        p = store.get(EntityType.PROCESS, "remove");
+        Assert.assertNull(p);
+    }
 
-  @Test
-  public void testSearch() throws Exception {
-    //TODO
-  }
+    @Test
+    public void testSearch() throws Exception {
+        //TODO
+    }
 
-  @BeforeSuite
-  @AfterSuite
-  public void cleanup() throws IOException {
-    Path path = new Path(StartupProperties.get().
-        getProperty("config.store.uri"));
-    FileSystem fs = FileSystem.get(path.toUri(), new Configuration());
-    fs.delete(path, true);
-    LOG.info("Cleaned up " + path);
-  }
+    @BeforeSuite
+    @AfterSuite
+    public void cleanup() throws IOException {
+        Path path = new Path(StartupProperties.get().
+                getProperty("config.store.uri"));
+        FileSystem fs = FileSystem.get(path.toUri(), new Configuration());
+        fs.delete(path, true);
+        LOG.info("Cleaned up " + path);
+    }
 }

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/entity/v0/EntityGraphTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/entity/v0/EntityGraphTest.java b/common/src/test/java/org/apache/falcon/entity/v0/EntityGraphTest.java
index 6078e01..9b4bffc 100644
--- a/common/src/test/java/org/apache/falcon/entity/v0/EntityGraphTest.java
+++ b/common/src/test/java/org/apache/falcon/entity/v0/EntityGraphTest.java
@@ -18,22 +18,19 @@
 
 package org.apache.falcon.entity.v0;
 
-import java.util.Set;
-
 import org.apache.falcon.entity.AbstractTestBase;
 import org.apache.falcon.entity.store.ConfigurationStore;
 import org.apache.falcon.entity.v0.cluster.Cluster;
 import org.apache.falcon.entity.v0.feed.Clusters;
 import org.apache.falcon.entity.v0.feed.Feed;
-import org.apache.falcon.entity.v0.process.Input;
-import org.apache.falcon.entity.v0.process.Inputs;
-import org.apache.falcon.entity.v0.process.Output;
-import org.apache.falcon.entity.v0.process.Outputs;
+import org.apache.falcon.entity.v0.process.*;
 import org.apache.falcon.entity.v0.process.Process;
 import org.testng.Assert;
 import org.testng.annotations.Test;
 
-public class EntityGraphTest extends AbstractTestBase{
+import java.util.Set;
+
+public class EntityGraphTest extends AbstractTestBase {
 
     private ConfigurationStore store = ConfigurationStore.get();
 
@@ -101,7 +98,9 @@ public class EntityGraphTest extends AbstractTestBase{
     }
 
     private Feed addInput(Process process, String feed, Cluster cluster) {
-        if (process.getInputs() == null) process.setInputs(new Inputs());
+        if (process.getInputs() == null) {
+            process.setInputs(new Inputs());
+        }
         Inputs inputs = process.getInputs();
         Input input = new Input();
         input.setFeed(feed);
@@ -114,11 +113,13 @@ public class EntityGraphTest extends AbstractTestBase{
                 new org.apache.falcon.entity.v0.feed.Cluster();
         feedCluster.setName(cluster.getName());
         clusters.getClusters().add(feedCluster);
-        return f1; 
+        return f1;
     }
 
     private void attachInput(Process process, Feed feed) {
-        if (process.getInputs() == null) process.setInputs(new Inputs());
+        if (process.getInputs() == null) {
+            process.setInputs(new Inputs());
+        }
         Inputs inputs = process.getInputs();
         Input input = new Input();
         input.setFeed(feed.getName());
@@ -126,7 +127,9 @@ public class EntityGraphTest extends AbstractTestBase{
     }
 
     private Feed addOutput(Process process, String feed, Cluster cluster) {
-        if (process.getOutputs() == null) process.setOutputs(new Outputs());
+        if (process.getOutputs() == null) {
+            process.setOutputs(new Outputs());
+        }
         Outputs Outputs = process.getOutputs();
         Output Output = new Output();
         Output.setFeed(feed);

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/group/FeedGroupMapTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/group/FeedGroupMapTest.java b/common/src/test/java/org/apache/falcon/group/FeedGroupMapTest.java
index bd5fedb..f7812b9 100644
--- a/common/src/test/java/org/apache/falcon/group/FeedGroupMapTest.java
+++ b/common/src/test/java/org/apache/falcon/group/FeedGroupMapTest.java
@@ -17,12 +17,7 @@
  */
 package org.apache.falcon.group;
 
-import java.util.Map;
-
-import javax.xml.bind.JAXBException;
-
 import junit.framework.Assert;
-
 import org.apache.falcon.FalconException;
 import org.apache.falcon.entity.AbstractTestBase;
 import org.apache.falcon.entity.store.ConfigurationStore;
@@ -36,166 +31,169 @@ import org.testng.annotations.BeforeClass;
 import org.testng.annotations.BeforeMethod;
 import org.testng.annotations.Test;
 
-public class FeedGroupMapTest extends AbstractTestBase{
-	private ConfigurationStore store = ConfigurationStore.get();
-	private static Cluster cluster;
-
-	@BeforeClass
-	public void setUp() throws Exception {
-		cluster = (Cluster) EntityType.CLUSTER
-				.getUnmarshaller()
-				.unmarshal(
-						FeedGroupMapTest.class
-								.getResourceAsStream("/config/cluster/cluster-0.1.xml"));
-	}
-
-	@BeforeMethod
-	public void cleanup() throws Exception {
-	    cleanupStore();
-	}
-	
-	@Test
-	public void testOnAdd() throws FalconException, JAXBException {
-		store.publish(EntityType.CLUSTER, cluster);
-		Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
-				FeedGroupMapTest.class
-						.getResourceAsStream("/config/feed/feed-0.1.xml"));
-		feed1.setName("f1");
-		feed1.setGroups("group1,group2,group3");
-		Location location = new Location();
-		location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
-		location.setType(LocationType.DATA);
-		feed1.setLocations(new Locations());
-		feed1.getLocations().getLocations().add(location);
-		store.publish(EntityType.FEED, feed1);
-		Map<String, FeedGroup> groupMapping = FeedGroupMap.get()
-				.getGroupsMapping();
-
-		FeedGroup group = groupMapping.get("group1");
-		Assert.assertEquals(group.getName(), "group1");
-		Assert.assertEquals(group.getFeeds().size(), 1);
-		assertFields(group, feed1);
-
-		group = groupMapping.get("group2");
-		Assert.assertEquals(group.getName(), "group2");
-		Assert.assertEquals(group.getFeeds().size(), 1);
-		assertFields(group, feed1);
-
-		group = groupMapping.get("group3");
-		Assert.assertEquals(group.getName(), "group3");
-		Assert.assertEquals(group.getFeeds().size(), 1);
-		assertFields(group, feed1);
-
-		Feed feed2 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
-				FeedGroupMapTest.class
-						.getResourceAsStream("/config/feed/feed-0.1.xml"));
-
-		feed2.setName("f2");
-		feed2.setGroups("group1,group5,group3");
-		location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/${MONTH}/${DAY}/ad2");
+import javax.xml.bind.JAXBException;
+import java.util.Map;
+
+public class FeedGroupMapTest extends AbstractTestBase {
+    private ConfigurationStore store = ConfigurationStore.get();
+    private static Cluster cluster;
+
+    @BeforeClass
+    public void setUp() throws Exception {
+        cluster = (Cluster) EntityType.CLUSTER
+                .getUnmarshaller()
+                .unmarshal(
+                        FeedGroupMapTest.class
+                                .getResourceAsStream("/config/cluster/cluster-0.1.xml"));
+    }
+
+    @BeforeMethod
+    public void cleanup() throws Exception {
+        cleanupStore();
+    }
+
+    @Test
+    public void testOnAdd() throws FalconException, JAXBException {
+        store.publish(EntityType.CLUSTER, cluster);
+        Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
+                FeedGroupMapTest.class
+                        .getResourceAsStream("/config/feed/feed-0.1.xml"));
+        feed1.setName("f1");
+        feed1.setGroups("group1,group2,group3");
+        Location location = new Location();
+        location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
+        location.setType(LocationType.DATA);
+        feed1.setLocations(new Locations());
+        feed1.getLocations().getLocations().add(location);
+        store.publish(EntityType.FEED, feed1);
+        Map<String, FeedGroup> groupMapping = FeedGroupMap.get()
+                .getGroupsMapping();
+
+        FeedGroup group = groupMapping.get("group1");
+        Assert.assertEquals(group.getName(), "group1");
+        Assert.assertEquals(group.getFeeds().size(), 1);
+        assertFields(group, feed1);
+
+        group = groupMapping.get("group2");
+        Assert.assertEquals(group.getName(), "group2");
+        Assert.assertEquals(group.getFeeds().size(), 1);
+        assertFields(group, feed1);
+
+        group = groupMapping.get("group3");
+        Assert.assertEquals(group.getName(), "group3");
+        Assert.assertEquals(group.getFeeds().size(), 1);
+        assertFields(group, feed1);
+
+        Feed feed2 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
+                FeedGroupMapTest.class
+                        .getResourceAsStream("/config/feed/feed-0.1.xml"));
+
+        feed2.setName("f2");
+        feed2.setGroups("group1,group5,group3");
+        location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/${MONTH}/${DAY}/ad2");
         location.setType(LocationType.DATA);
         feed2.setLocations(new Locations());
         feed2.getLocations().getLocations().add(location);
-		store.publish(EntityType.FEED, feed2);
-		groupMapping = FeedGroupMap.get().getGroupsMapping();
-
-		group = groupMapping.get("group1");
-		Assert.assertEquals(group.getName(), "group1");
-		Assert.assertEquals(group.getFeeds().size(), 2);
-		assertFields(group, feed2);
-
-		group = groupMapping.get("group2");
-		Assert.assertEquals(group.getName(), "group2");
-		Assert.assertEquals(group.getFeeds().size(), 1);
-		assertFields(group, feed2);
-
-		group = groupMapping.get("group3");
-		Assert.assertEquals(group.getName(), "group3");
-		Assert.assertEquals(group.getFeeds().size(), 2);
-		assertFields(group, feed2);
-
-		group = groupMapping.get("group5");
-		Assert.assertEquals(group.getName(), "group5");
-		Assert.assertEquals(group.getFeeds().size(), 1);
-		assertFields(group, feed2);
-
-	}
-
-	@Test
-	public void testOnRemove() throws FalconException, JAXBException {
-		Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
-				FeedGroupMapTest.class
-						.getResourceAsStream("/config/feed/feed-0.1.xml"));
-		feed1.setName("f1");
-		store.publish(EntityType.CLUSTER, cluster);
-		feed1.setGroups("group7,group8,group9");
-		Location location = new Location();
-		location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
+        store.publish(EntityType.FEED, feed2);
+        groupMapping = FeedGroupMap.get().getGroupsMapping();
+
+        group = groupMapping.get("group1");
+        Assert.assertEquals(group.getName(), "group1");
+        Assert.assertEquals(group.getFeeds().size(), 2);
+        assertFields(group, feed2);
+
+        group = groupMapping.get("group2");
+        Assert.assertEquals(group.getName(), "group2");
+        Assert.assertEquals(group.getFeeds().size(), 1);
+        assertFields(group, feed2);
+
+        group = groupMapping.get("group3");
+        Assert.assertEquals(group.getName(), "group3");
+        Assert.assertEquals(group.getFeeds().size(), 2);
+        assertFields(group, feed2);
+
+        group = groupMapping.get("group5");
+        Assert.assertEquals(group.getName(), "group5");
+        Assert.assertEquals(group.getFeeds().size(), 1);
+        assertFields(group, feed2);
+
+    }
+
+    @Test
+    public void testOnRemove() throws FalconException, JAXBException {
+        Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
+                FeedGroupMapTest.class
+                        .getResourceAsStream("/config/feed/feed-0.1.xml"));
+        feed1.setName("f1");
+        store.publish(EntityType.CLUSTER, cluster);
+        feed1.setGroups("group7,group8,group9");
+        Location location = new Location();
+        location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
         location.setType(LocationType.DATA);
         feed1.setLocations(new Locations());
         feed1.getLocations().getLocations().add(location);
-		store.publish(EntityType.FEED, feed1);
-
-		Feed feed2 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
-				FeedGroupMapTest.class
-						.getResourceAsStream("/config/feed/feed-0.1.xml"));
-		feed2.setName("f2");
-		feed2.setGroups("group7,group8,group10");
-		location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/${MONTH}/${DAY}/ad2");
+        store.publish(EntityType.FEED, feed1);
+
+        Feed feed2 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
+                FeedGroupMapTest.class
+                        .getResourceAsStream("/config/feed/feed-0.1.xml"));
+        feed2.setName("f2");
+        feed2.setGroups("group7,group8,group10");
+        location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/${MONTH}/${DAY}/ad2");
         location.setType(LocationType.DATA);
         feed2.setLocations(new Locations());
         feed2.getLocations().getLocations().add(location);
-		store.publish(EntityType.FEED, feed2);
+        store.publish(EntityType.FEED, feed2);
 
-		Map<String, FeedGroup> groupMapping = FeedGroupMap.get()
-				.getGroupsMapping();
+        Map<String, FeedGroup> groupMapping = FeedGroupMap.get()
+                .getGroupsMapping();
 
-		store.remove(EntityType.FEED, "f2");
+        store.remove(EntityType.FEED, "f2");
 
-		FeedGroup group = groupMapping.get("group7");
-		Assert.assertEquals(group.getName(), "group7");
-		Assert.assertEquals(group.getFeeds().size(), 1);
+        FeedGroup group = groupMapping.get("group7");
+        Assert.assertEquals(group.getName(), "group7");
+        Assert.assertEquals(group.getFeeds().size(), 1);
 
-		group = groupMapping.get("group8");
-		Assert.assertEquals(group.getName(), "group8");
-		Assert.assertEquals(group.getFeeds().size(), 1);
+        group = groupMapping.get("group8");
+        Assert.assertEquals(group.getName(), "group8");
+        Assert.assertEquals(group.getFeeds().size(), 1);
 
-		group = groupMapping.get("group10");
-		Assert.assertEquals(null, group);
+        group = groupMapping.get("group10");
+        Assert.assertEquals(null, group);
 
-		store.remove(EntityType.FEED, "f1");
+        store.remove(EntityType.FEED, "f1");
 
-		group = groupMapping.get("group7");
-		Assert.assertEquals(null, group);
+        group = groupMapping.get("group7");
+        Assert.assertEquals(null, group);
 
-		group = groupMapping.get("group8");
-		Assert.assertEquals(null, group);
+        group = groupMapping.get("group8");
+        Assert.assertEquals(null, group);
 
-		group = groupMapping.get("group9");
-		Assert.assertEquals(null, group);
+        group = groupMapping.get("group9");
+        Assert.assertEquals(null, group);
 
-	}
+    }
 
-	@Test
-	public void testNullGroup() throws FalconException, JAXBException {
-		Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
-				FeedGroupMapTest.class
-						.getResourceAsStream("/config/feed/feed-0.1.xml"));
-		feed1.setName("f5" + System.currentTimeMillis());
-		store.publish(EntityType.CLUSTER, cluster);
-		feed1.setGroups(null);
-		Location location = new Location();
-		location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
+    @Test
+    public void testNullGroup() throws FalconException, JAXBException {
+        Feed feed1 = (Feed) EntityType.FEED.getUnmarshaller().unmarshal(
+                FeedGroupMapTest.class
+                        .getResourceAsStream("/config/feed/feed-0.1.xml"));
+        feed1.setName("f5" + System.currentTimeMillis());
+        store.publish(EntityType.CLUSTER, cluster);
+        feed1.setGroups(null);
+        Location location = new Location();
+        location.setPath("/projects/bi/rmc/daily/ad/${YEAR}/fraud/${MONTH}-${DAY}/ad");
         location.setType(LocationType.DATA);
         feed1.setLocations(new Locations());
         feed1.getLocations().getLocations().add(location);
-		store.publish(EntityType.FEED, feed1);
+        store.publish(EntityType.FEED, feed1);
 
-	}
+    }
 
-	private void assertFields(FeedGroup group, Feed feed) {
-		Assert.assertEquals(group.getFrequency(), feed.getFrequency());
-		Assert.assertEquals(group.getDatePattern(),
-				"[${DAY}, ${MONTH}, ${YEAR}]");
-	}
+    private void assertFields(FeedGroup group, Feed feed) {
+        Assert.assertEquals(group.getFrequency(), feed.getFrequency());
+        Assert.assertEquals(group.getDatePattern(),
+                "[${DAY}, ${MONTH}, ${YEAR}]");
+    }
 }

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/security/CurrentUserTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/security/CurrentUserTest.java b/common/src/test/java/org/apache/falcon/security/CurrentUserTest.java
index 3ba095f..b31e3cc 100644
--- a/common/src/test/java/org/apache/falcon/security/CurrentUserTest.java
+++ b/common/src/test/java/org/apache/falcon/security/CurrentUserTest.java
@@ -23,7 +23,7 @@ import org.testng.annotations.Test;
 
 public class CurrentUserTest {
 
-    @Test(threadPoolSize = 10, invocationCount = 10,  timeOut = 10000)
+    @Test(threadPoolSize = 10, invocationCount = 10, timeOut = 10000)
     public void testGetUser() throws Exception {
         String id = Long.toString(System.nanoTime());
         CurrentUser.authenticate(id);

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/update/UpdateHelperTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/update/UpdateHelperTest.java b/common/src/test/java/org/apache/falcon/update/UpdateHelperTest.java
index 9777088..11e5d06 100644
--- a/common/src/test/java/org/apache/falcon/update/UpdateHelperTest.java
+++ b/common/src/test/java/org/apache/falcon/update/UpdateHelperTest.java
@@ -27,11 +27,7 @@ import org.apache.falcon.entity.parser.ProcessEntityParser;
 import org.apache.falcon.entity.v0.EntityType;
 import org.apache.falcon.entity.v0.Frequency;
 import org.apache.falcon.entity.v0.SchemaHelper;
-import org.apache.falcon.entity.v0.feed.Feed;
-import org.apache.falcon.entity.v0.feed.LocationType;
-import org.apache.falcon.entity.v0.feed.Partition;
-import org.apache.falcon.entity.v0.feed.Properties;
-import org.apache.falcon.entity.v0.feed.Property;
+import org.apache.falcon.entity.v0.feed.*;
 import org.apache.falcon.entity.v0.process.PolicyType;
 import org.apache.falcon.entity.v0.process.Process;
 import org.testng.Assert;
@@ -53,40 +49,40 @@ public class UpdateHelperTest extends AbstractTestBase {
         setup();
     }
 
-	@AfterClass
-	public void tearDown() {
-		this.dfsCluster.shutdown();
-	}
+    @AfterClass
+    public void tearDown() {
+        this.dfsCluster.shutdown();
+    }
 
-	@BeforeMethod
-	public void setUp() throws Exception {
+    @BeforeMethod
+    public void setUp() throws Exception {
         storeEntity(EntityType.CLUSTER, "testCluster");
         storeEntity(EntityType.CLUSTER, "backupCluster");
         storeEntity(EntityType.FEED, "clicksFeed");
         storeEntity(EntityType.FEED, "impressionFeed");
         storeEntity(EntityType.FEED, "imp-click-join1");
         storeEntity(EntityType.FEED, "imp-click-join2");
-	}
-	
-	@Test
-	public void testShouldUpdate2() throws Exception {
+    }
+
+    @Test
+    public void testShouldUpdate2() throws Exception {
         Feed oldFeed = parser.parseAndValidate(this.getClass()
                 .getResourceAsStream(FEED_XML));
         String cluster = "testCluster";
-        Feed newFeed = (Feed)oldFeed.clone();
+        Feed newFeed = (Feed) oldFeed.clone();
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldFeed, newFeed, cluster));
-        
+
         newFeed.setGroups("newgroups");
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldFeed, newFeed, cluster));
         newFeed.getLateArrival().setCutOff(Frequency.fromString("hours(8)"));
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldFeed, newFeed, cluster));
         newFeed.setFrequency(Frequency.fromString("days(1)"));
         Assert.assertTrue(UpdateHelper.shouldUpdate(oldFeed, newFeed, cluster));
-        
+
         Process oldProcess = processParser.parseAndValidate(this.getClass().
                 getResourceAsStream(PROCESS_XML));
         Process newProcess = (Process) oldProcess.clone();
-        
+
         newProcess.getRetry().setPolicy(PolicyType.FINAL);
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldProcess, newProcess, cluster));
         newProcess.getLateProcess().getLateInputs().remove(1);
@@ -94,15 +90,15 @@ public class UpdateHelperTest extends AbstractTestBase {
         newProcess.getLateProcess().setPolicy(PolicyType.PERIODIC);
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldProcess, newProcess, cluster));
         newProcess.setFrequency(Frequency.fromString("days(1)"));
-        Assert.assertTrue(UpdateHelper.shouldUpdate(oldProcess, newProcess, cluster));        
-	}
-	
+        Assert.assertTrue(UpdateHelper.shouldUpdate(oldProcess, newProcess, cluster));
+    }
+
     @Test
     public void testShouldUpdate() throws Exception {
         Feed oldFeed = parser.parseAndValidate(this.getClass()
                 .getResourceAsStream(FEED_XML));
 
-        Feed newFeed = (Feed)oldFeed.clone();
+        Feed newFeed = (Feed) oldFeed.clone();
         Process process = processParser.parseAndValidate(this.getClass().
                 getResourceAsStream(PROCESS_XML));
 
@@ -142,11 +138,13 @@ public class UpdateHelperTest extends AbstractTestBase {
         newFeed.getProperties().getProperties().remove(0);
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldFeed, newFeed, process));
 
-        FeedHelper.getCluster(newFeed, process.getClusters().getClusters().get(0).getName()).getValidity().setStart(SchemaHelper.parseDateUTC("2012-11-01T00:00Z"));
+        FeedHelper.getCluster(newFeed, process.getClusters().getClusters().get(0).getName()).getValidity().setStart(
+                SchemaHelper.parseDateUTC("2012-11-01T00:00Z"));
         Assert.assertTrue(UpdateHelper.shouldUpdate(oldFeed, newFeed, process));
 
         FeedHelper.getCluster(newFeed, process.getClusters().getClusters().get(0).getName()).getValidity().
-                setStart(FeedHelper.getCluster(oldFeed, process.getClusters().getClusters().get(0).getName()).getValidity().getStart());
+                setStart(FeedHelper.getCluster(oldFeed,
+                        process.getClusters().getClusters().get(0).getName()).getValidity().getStart());
         Assert.assertFalse(UpdateHelper.shouldUpdate(oldFeed, newFeed, process));
     }
 }

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/java/org/apache/falcon/util/StartupPropertiesTest.java
----------------------------------------------------------------------
diff --git a/common/src/test/java/org/apache/falcon/util/StartupPropertiesTest.java b/common/src/test/java/org/apache/falcon/util/StartupPropertiesTest.java
index 6e6a782..fc3d604 100644
--- a/common/src/test/java/org/apache/falcon/util/StartupPropertiesTest.java
+++ b/common/src/test/java/org/apache/falcon/util/StartupPropertiesTest.java
@@ -18,17 +18,18 @@
 
 package org.apache.falcon.util;
 
-import static org.testng.AssertJUnit.assertEquals;
 import org.testng.annotations.BeforeClass;
 import org.testng.annotations.Test;
 
+import static org.testng.AssertJUnit.assertEquals;
+
 @Test
 public class StartupPropertiesTest {
     @BeforeClass
     public void setUp() {
         StartupProperties.get();
     }
-    
+
     public void testDomain() {
         StartupProperties props = (StartupProperties) StartupProperties.get();
         assertEquals("debug", props.getDomain());

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/cluster/cluster-0.1.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/cluster/cluster-0.1.xml b/common/src/test/resources/config/cluster/cluster-0.1.xml
index ee59616..cf53398 100644
--- a/common/src/test/resources/config/cluster/cluster-0.1.xml
+++ b/common/src/test/resources/config/cluster/cluster-0.1.xml
@@ -17,26 +17,27 @@
   limitations under the License.
   -->
 
-<cluster colo="default" description="" name="testCluster" xmlns="uri:falcon:cluster:0.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
-	<interfaces>
-		<interface type="readonly" endpoint="hftp://localhost:50010"
-			version="0.20.2" />
-		<interface type="write" endpoint="hdfs://localhost:8020"
-			version="0.20.2" />
-		<interface type="execute" endpoint="localhost:8021" version="0.20.2" />
-		<interface type="workflow" endpoint="http://localhost:11000/oozie/"
-			version="3.1" />
-		<interface type="messaging" endpoint="tcp://localhost:61616?daemon=true"
-			version="5.1.6" />
-		<interface type="registry" endpoint="Hcat" version="1" />
-	</interfaces>
-	<locations>
-		<location name="staging" path="/projects/falcon/staging" />
-		<location name="temp" path="/tmp" />
-		<location name="working" path="/projects/falcon/working" />
-	</locations>
-	<properties>
-		<property name="field1" value="value1" />
-		<property name="field2" value="value2" />
-	</properties>
+<cluster colo="default" description="" name="testCluster" xmlns="uri:falcon:cluster:0.1"
+        >
+    <interfaces>
+        <interface type="readonly" endpoint="hftp://localhost:50010"
+                   version="0.20.2"/>
+        <interface type="write" endpoint="hdfs://localhost:8020"
+                   version="0.20.2"/>
+        <interface type="execute" endpoint="localhost:8021" version="0.20.2"/>
+        <interface type="workflow" endpoint="http://localhost:11000/oozie/"
+                   version="3.1"/>
+        <interface type="messaging" endpoint="tcp://localhost:61616?daemon=true"
+                   version="5.1.6"/>
+        <interface type="registry" endpoint="Hcat" version="1"/>
+    </interfaces>
+    <locations>
+        <location name="staging" path="/projects/falcon/staging"/>
+        <location name="temp" path="/tmp"/>
+        <location name="working" path="/projects/falcon/working"/>
+    </locations>
+    <properties>
+        <property name="field1" value="value1"/>
+        <property name="field2" value="value2"/>
+    </properties>
 </cluster>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/feed/feed-0.1.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/feed/feed-0.1.xml b/common/src/test/resources/config/feed/feed-0.1.xml
index ee04499..0269fa0 100644
--- a/common/src/test/resources/config/feed/feed-0.1.xml
+++ b/common/src/test/resources/config/feed/feed-0.1.xml
@@ -16,10 +16,11 @@
   See the License for the specific language governing permissions and
   limitations under the License.
   -->
-<feed description="clicks log" name="clicks" xmlns="uri:falcon:feed:0.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
-   <partitions>
-        <partition name="fraud" />
-        <partition name="good" />
+<feed description="clicks log" name="clicks" xmlns="uri:falcon:feed:0.1"
+        >
+    <partitions>
+        <partition name="fraud"/>
+        <partition name="good"/>
     </partitions>
 
     <groups>online,bi</groups>
@@ -27,31 +28,33 @@
 
     <frequency>hours(1)</frequency>
     <timezone>UTC</timezone>
-    
-    <late-arrival cut-off="hours(6)" />
+
+    <late-arrival cut-off="hours(6)"/>
 
     <clusters>
         <cluster name="testCluster" type="source">
             <validity start="2011-11-01T00:00Z" end="2011-12-31T00:00Z"/>
-            <retention limit="hours(48)" action="delete" /> <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
-       		<locations>
-        		<location type="data" path="/projects/falcon/clicks" />
-        		<location type="stats" path="/projects/falcon/clicksStats" />
-        		<location type="meta" path="/projects/falcon/clicksMetaData" />
-    		</locations>
+            <retention limit="hours(48)" action="delete"/>
+            <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
+            <locations>
+                <location type="data" path="/projects/falcon/clicks"/>
+                <location type="stats" path="/projects/falcon/clicksStats"/>
+                <location type="meta" path="/projects/falcon/clicksMetaData"/>
+            </locations>
         </cluster>
         <cluster name="backupCluster" type="target">
             <validity start="2011-11-01T00:00Z" end="2011-12-31T00:00Z"/>
-            <retention limit="hours(6)" action="archive" /> <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
+            <retention limit="hours(6)" action="archive"/>
+            <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
         </cluster>
     </clusters>
 
     <locations>
-        <location type="data" path="/projects/falcon/clicks" />
-        <location type="stats" path="/projects/falcon/clicksStats" />
-        <location type="meta" path="/projects/falcon/clicksMetaData" />
+        <location type="data" path="/projects/falcon/clicks"/>
+        <location type="stats" path="/projects/falcon/clicksStats"/>
+        <location type="meta" path="/projects/falcon/clicksMetaData"/>
     </locations>
-    
-    <ACL owner="testuser" group="group" permission="0x755" />
-    <schema location="/schema/clicks" provider="protobuf" />
+
+    <ACL owner="testuser" group="group" permission="0x755"/>
+    <schema location="/schema/clicks" provider="protobuf"/>
 </feed>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/feed/feed-0.2.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/feed/feed-0.2.xml b/common/src/test/resources/config/feed/feed-0.2.xml
index 1f15cdf..a3f4f4d 100644
--- a/common/src/test/resources/config/feed/feed-0.2.xml
+++ b/common/src/test/resources/config/feed/feed-0.2.xml
@@ -16,35 +16,38 @@
   See the License for the specific language governing permissions and
   limitations under the License.
   -->
-<feed description="clicks log" name="clicks" xmlns="uri:falcon:feed:0.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
-   <partitions>
-        <partition name="fraud" />
-        <partition name="good" />
+<feed description="clicks log" name="clicks" xmlns="uri:falcon:feed:0.1"
+        >
+    <partitions>
+        <partition name="fraud"/>
+        <partition name="good"/>
     </partitions>
 
     <groups>online,bi</groups>
 
     <frequency>hours(1)</frequency>
     <timezone>UTC</timezone>
-    <late-arrival cut-off="hours(6)" />
+    <late-arrival cut-off="hours(6)"/>
 
     <clusters>
         <cluster name="testCluster" type="source" partition="*/${cluster.colo}">
             <validity start="2021-11-01T00:00Z" end="2021-12-31T00:00Z"/>
-            <retention limit="hours(48)" action="delete" /> <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
+            <retention limit="hours(48)" action="delete"/>
+            <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
         </cluster>
         <cluster name="backupCluster" type="target">
             <validity start="2011-11-01T00:00Z" end="2011-12-31T00:00Z"/>
-            <retention limit="hours(6)" action="archive" /> <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
+            <retention limit="hours(6)" action="archive"/>
+            <!-- Limit can be in Time or Instances 100, Action ENUM DELETE,ARCHIVE -->
         </cluster>
     </clusters>
 
     <locations>
-        <location type="data" path="/projects/falcon/clicks" />
-        <location type="stats" path="/projects/falcon/clicksStats" />
-        <location type="meta" path="/projects/falcon/clicksMetaData" />
+        <location type="data" path="/projects/falcon/clicks"/>
+        <location type="stats" path="/projects/falcon/clicksStats"/>
+        <location type="meta" path="/projects/falcon/clicksMetaData"/>
     </locations>
 
-    <ACL owner="testuser" group="group" permission="0x755" />
-    <schema location="/schema/clicks" provider="protobuf" />
+    <ACL owner="testuser" group="group" permission="0x755"/>
+    <schema location="/schema/clicks" provider="protobuf"/>
 </feed>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/process/process-0.1.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/process/process-0.1.xml b/common/src/test/resources/config/process/process-0.1.xml
index 196f2cb..bb5cd35 100644
--- a/common/src/test/resources/config/process/process-0.1.xml
+++ b/common/src/test/resources/config/process/process-0.1.xml
@@ -16,10 +16,10 @@
   See the License for the specific language governing permissions and
   limitations under the License.
   -->
-<process name="sample" xmlns="uri:falcon:process:0.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
+<process name="sample" xmlns="uri:falcon:process:0.1">
     <clusters>
         <cluster name="testCluster">
-            <validity start="2011-11-02T00:00Z" end="2011-12-30T00:00Z" />
+            <validity start="2011-11-02T00:00Z" end="2011-12-30T00:00Z"/>
         </cluster>
     </clusters>
     <parallel>1</parallel>
@@ -28,27 +28,27 @@
 
     <!-- what -->
     <inputs>
-        <input name="impression" feed="impressionFeed" start="today(0,0)" end="today(2,0)" partition="*/US" />
-        <input name="clicks" feed="clicksFeed" start="yesterday(0,0)" end="yesterday(20,0)" />
+        <input name="impression" feed="impressionFeed" start="today(0,0)" end="today(2,0)" partition="*/US"/>
+        <input name="clicks" feed="clicksFeed" start="yesterday(0,0)" end="yesterday(20,0)"/>
     </inputs>
 
     <outputs>
-        <output name="impOutput" feed="imp-click-join1" instance="today(0,0)" />
-        <output name="clicksOutput" feed="imp-click-join2" instance="today(0,0)" />
+        <output name="impOutput" feed="imp-click-join1" instance="today(0,0)"/>
+        <output name="clicksOutput" feed="imp-click-join2" instance="today(0,0)"/>
     </outputs>
 
     <!-- how -->
     <properties>
-        <property name="name1" value="value1" />
-        <property name="name2" value="value2" />
+        <property name="name1" value="value1"/>
+        <property name="name2" value="value2"/>
     </properties>
 
-    <workflow engine="oozie" path="/path/to/workflow" />
+    <workflow engine="oozie" path="/path/to/workflow"/>
 
-    <retry policy="periodic" delay="minutes(10)" attempts="3" />
+    <retry policy="periodic" delay="minutes(10)" attempts="3"/>
 
     <late-process policy="exp-backoff" delay="hours(1)">
-        <late-input input="impression" workflow-path="himpression/late/workflow" />
-        <late-input input="clicks" workflow-path="hdfs://clicks/late/workflow" />
+        <late-input input="impression" workflow-path="himpression/late/workflow"/>
+        <late-input input="clicks" workflow-path="hdfs://clicks/late/workflow"/>
     </late-process>
 </process>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/process/process-0.2.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/process/process-0.2.xml b/common/src/test/resources/config/process/process-0.2.xml
index 7e817e3..c4cd83e 100644
--- a/common/src/test/resources/config/process/process-0.2.xml
+++ b/common/src/test/resources/config/process/process-0.2.xml
@@ -16,43 +16,43 @@
   See the License for the specific language governing permissions and
   limitations under the License.
   -->
-<process name="sample" xmlns="uri:falcon:process:0.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
+<process name="sample" xmlns="uri:falcon:process:0.1">
     <clusters>
         <cluster name="testCluster">
-            <validity start="2011-11-02T00:00Z" end="2011-12-30T00:00Z" />
+            <validity start="2011-11-02T00:00Z" end="2011-12-30T00:00Z"/>
         </cluster>
     </clusters>
-	<!-- when -->
-	<parallel>1</parallel>
-	<order>LIFO</order>
-	<frequency>hours(1)</frequency>
+    <!-- when -->
+    <parallel>1</parallel>
+    <order>LIFO</order>
+    <frequency>hours(1)</frequency>
     <timezone>UTC</timezone>
 
-	<!-- what -->
-	<inputs>
-		<input name="impression" feed="impressionFeed" start="today(0,0)"
-			end="today(2,0)" partition="*/US" />
-		<input name="clicks" feed="clicks" start="yesterday(0,0)"
-			end="yesterday(20,0)" />
-	</inputs>
-
-	<outputs>
-		<output name="impOutput" feed="imp-click-join1" instance="today(0,0)" />
-		<output name="clicksOutput" feed="imp-click-join2" instance="today(0,0)" />
-	</outputs>
-
-	<!-- how -->
-	<properties>
-		<property name="name1" value="value1" />
-		<property name="name2" value="value2" />
-	</properties>
-
-	<workflow engine="oozie" path="/path/to/workflow" />
-
-	<retry policy="periodic" delay="minutes(10)" attempts="3" />
-
-	<late-process policy="exp-backoff" delay="hours(1)">
-		<late-input input="impression" workflow-path="himpression/late/workflow" />
-		<late-input input="clicks" workflow-path="hdfs://clicks/late/workflow" />
-	</late-process>
+    <!-- what -->
+    <inputs>
+        <input name="impression" feed="impressionFeed" start="today(0,0)"
+               end="today(2,0)" partition="*/US"/>
+        <input name="clicks" feed="clicks" start="yesterday(0,0)"
+               end="yesterday(20,0)"/>
+    </inputs>
+
+    <outputs>
+        <output name="impOutput" feed="imp-click-join1" instance="today(0,0)"/>
+        <output name="clicksOutput" feed="imp-click-join2" instance="today(0,0)"/>
+    </outputs>
+
+    <!-- how -->
+    <properties>
+        <property name="name1" value="value1"/>
+        <property name="name2" value="value2"/>
+    </properties>
+
+    <workflow engine="oozie" path="/path/to/workflow"/>
+
+    <retry policy="periodic" delay="minutes(10)" attempts="3"/>
+
+    <late-process policy="exp-backoff" delay="hours(1)">
+        <late-input input="impression" workflow-path="himpression/late/workflow"/>
+        <late-input input="clicks" workflow-path="hdfs://clicks/late/workflow"/>
+    </late-process>
 </process>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/common/src/test/resources/config/process/process-invalid.xml
----------------------------------------------------------------------
diff --git a/common/src/test/resources/config/process/process-invalid.xml b/common/src/test/resources/config/process/process-invalid.xml
index edd9311..d98a6e9 100644
--- a/common/src/test/resources/config/process/process-invalid.xml
+++ b/common/src/test/resources/config/process/process-invalid.xml
@@ -16,23 +16,23 @@
   See the License for the specific language governing permissions and
   limitations under the License.
   -->
-<Process name="" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
-  <clusters>
-    <cluster name="">cluster</cluster>
-  </clusters>
-  <frequency>frequency</frequency>
-  <periodicity>periodicity</periodicity>
-  <validity end="" start="">validity</validity>
-  <inputs>
-    <somenode>somenode</somenode>
-    <input end="" feed="" start="">input</input>
-  </inputs>
-  <outputs>
-    <output feed="" instance="">output</output>
-  </outputs>
-  <workflow engine="" path="">workflow</workflow>
-  <retry attempts="" delay="" delayUnit="" policy="">retry</retry>
-  <late-process delay="" delayUnit="" policy="">
-    <late-input input="" workflow-path="">late-input</late-input>
-  </late-process>
+<Process name="">
+    <clusters>
+        <cluster name="">cluster</cluster>
+    </clusters>
+    <frequency>frequency</frequency>
+    <periodicity>periodicity</periodicity>
+    <validity end="" start="">validity</validity>
+    <inputs>
+        <somenode>somenode</somenode>
+        <input end="" feed="" start="">input</input>
+    </inputs>
+    <outputs>
+        <output feed="" instance="">output</output>
+    </outputs>
+    <workflow engine="" path="">workflow</workflow>
+    <retry attempts="" delay="" delayUnit="" policy="">retry</retry>
+    <late-process delay="" delayUnit="" policy="">
+        <late-input input="" workflow-path="">late-input</late-input>
+    </late-process>
 </Process>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/docs/pom.xml
----------------------------------------------------------------------
diff --git a/docs/pom.xml b/docs/pom.xml
index 8e8a444..a206267 100644
--- a/docs/pom.xml
+++ b/docs/pom.xml
@@ -16,7 +16,8 @@
   See the License for the specific language governing permissions and
   limitations under the License.
 -->
-<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
+<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
+         xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
     <modelVersion>4.0.0</modelVersion>
     <parent>
         <groupId>org.apache.falcon</groupId>
@@ -37,7 +38,6 @@
             <plugin>
                 <groupId>org.apache.maven.plugins</groupId>
                 <artifactId>maven-site-plugin</artifactId>
-                <version>3.0</version>
                 <dependencies>
                     <dependency>
                         <groupId>org.apache.maven.doxia</groupId>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/docs/src/site/site.xml
----------------------------------------------------------------------
diff --git a/docs/src/site/site.xml b/docs/src/site/site.xml
index cbc230b..4238dee 100644
--- a/docs/src/site/site.xml
+++ b/docs/src/site/site.xml
@@ -21,11 +21,11 @@
         <name>Falcon</name>
     </bannerLeft>
 
-  <skin>
-    <groupId>org.apache.maven.skins</groupId>
-    <artifactId>maven-classic-skin</artifactId>
-    <version>1.0</version>
-  </skin>
+    <skin>
+        <groupId>org.apache.maven.skins</groupId>
+        <artifactId>maven-classic-skin</artifactId>
+        <version>1.0</version>
+    </skin>
 
     <body/>
 

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/feed/src/main/java/org/apache/falcon/converter/OozieFeedMapper.java
----------------------------------------------------------------------
diff --git a/feed/src/main/java/org/apache/falcon/converter/OozieFeedMapper.java b/feed/src/main/java/org/apache/falcon/converter/OozieFeedMapper.java
index 73ce42b..d02f6c1 100644
--- a/feed/src/main/java/org/apache/falcon/converter/OozieFeedMapper.java
+++ b/feed/src/main/java/org/apache/falcon/converter/OozieFeedMapper.java
@@ -18,16 +18,7 @@
 
 package org.apache.falcon.converter;
 
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Date;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-
 import org.apache.commons.lang.StringUtils;
-import org.apache.hadoop.fs.Path;
 import org.apache.falcon.FalconException;
 import org.apache.falcon.Tag;
 import org.apache.falcon.entity.ClusterHelper;
@@ -50,8 +41,12 @@ import org.apache.falcon.oozie.coordinator.COORDINATORAPP;
 import org.apache.falcon.oozie.coordinator.SYNCDATASET;
 import org.apache.falcon.oozie.coordinator.WORKFLOW;
 import org.apache.falcon.oozie.workflow.WORKFLOWAPP;
+import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
+import java.io.IOException;
+import java.util.*;
+
 public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
 
     private static Logger LOG = Logger.getLogger(OozieFeedMapper.class);
@@ -61,8 +56,8 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
     private static final String RETENTION_WF_TEMPLATE = "/config/workflow/retention-workflow.xml";
     private static final String REPLICATION_COORD_TEMPLATE = "/config/coordinator/replication-coordinator.xml";
     private static final String REPLICATION_WF_TEMPLATE = "/config/workflow/replication-workflow.xml";
-    
-    private static final String FEED_PATH_SEP="#";
+
+    private static final String FEED_PATH_SEP = "#";
 
     public OozieFeedMapper(Feed feed) {
         super(feed);
@@ -86,7 +81,8 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
         org.apache.falcon.entity.v0.feed.Cluster feedCluster = FeedHelper.getCluster(feed, cluster.getName());
 
         if (feedCluster.getValidity().getEnd().before(new Date())) {
-            LOG.warn("Feed Retention is not applicable as Feed's end time for cluster " + cluster.getName() + " is not in the future");
+            LOG.warn("Feed Retention is not applicable as Feed's end time for cluster " + cluster.getName()
+                    + " is not in the future");
             return null;
         }
         COORDINATORAPP retentionApp = new COORDINATORAPP();
@@ -121,20 +117,20 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
             Map<String, String> props = createCoordDefaultConfiguration(cluster, wfPath, wfName);
 
             org.apache.falcon.entity.v0.feed.Cluster feedCluster = FeedHelper.getCluster(feed, cluster.getName());
-            String feedPathMask = getLocationURI(cluster, feed,LocationType.DATA);
-			String metaPathMask = getLocationURI(cluster, feed, LocationType.META);
+            String feedPathMask = getLocationURI(cluster, feed, LocationType.DATA);
+            String metaPathMask = getLocationURI(cluster, feed, LocationType.META);
             String statsPathMask = getLocationURI(cluster, feed, LocationType.STATS);
             String tmpPathMask = getLocationURI(cluster, feed, LocationType.TMP);
 
             StringBuilder feedBasePaths = new StringBuilder(feedPathMask);
-            if(metaPathMask!=null){
-            	feedBasePaths.append(FEED_PATH_SEP).append(metaPathMask);
+            if (metaPathMask != null) {
+                feedBasePaths.append(FEED_PATH_SEP).append(metaPathMask);
             }
-            if(statsPathMask!=null){
-            	feedBasePaths.append(FEED_PATH_SEP).append(statsPathMask);
+            if (statsPathMask != null) {
+                feedBasePaths.append(FEED_PATH_SEP).append(statsPathMask);
             }
-            if(tmpPathMask!=null){
-            	feedBasePaths.append(FEED_PATH_SEP).append(tmpPathMask);
+            if (tmpPathMask != null) {
+                feedBasePaths.append(FEED_PATH_SEP).append(tmpPathMask);
             }
 
             props.put("feedDataPath", feedBasePaths.toString().replaceAll("\\$\\{", "\\?\\{"));
@@ -153,23 +149,25 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
         }
     }
 
-    private List<COORDINATORAPP> getReplicationCoordinators(Cluster targetCluster, Path bundlePath) throws FalconException {
+    private List<COORDINATORAPP> getReplicationCoordinators(Cluster targetCluster, Path bundlePath)
+            throws FalconException {
         Feed feed = getEntity();
         List<COORDINATORAPP> replicationCoords = new ArrayList<COORDINATORAPP>();
-        
+
         if (FeedHelper.getCluster(feed, targetCluster.getName()).getType() == ClusterType.TARGET) {
             String coordName = EntityUtil.getWorkflowName(Tag.REPLICATION, feed).toString();
             Path basePath = getCoordPath(bundlePath, coordName);
             createReplicatonWorkflow(targetCluster, basePath, coordName);
-            
+
             for (org.apache.falcon.entity.v0.feed.Cluster feedCluster : feed.getClusters().getClusters()) {
                 if (feedCluster.getType() == ClusterType.SOURCE) {
                     COORDINATORAPP coord = createAndGetCoord(feed,
-                            (Cluster) ConfigurationStore.get().get(EntityType.CLUSTER, feedCluster.getName()), targetCluster,
+                            (Cluster) ConfigurationStore.get().get(EntityType.CLUSTER, feedCluster.getName()),
+                            targetCluster,
                             bundlePath);
-					if (coord != null) {
-						replicationCoords.add(coord);
-					}
+                    if (coord != null) {
+                        replicationCoords.add(coord);
+                    }
                 }
             }
 
@@ -183,14 +181,17 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
         String coordName;
         try {
             replicationCoord = getCoordinatorTemplate(REPLICATION_COORD_TEMPLATE);
-            coordName = EntityUtil.getWorkflowName(Tag.REPLICATION, Arrays.asList(srcCluster.getName()), feed).toString();
+            coordName = EntityUtil.getWorkflowName(Tag.REPLICATION, Arrays.asList(srcCluster.getName()),
+                    feed).toString();
             replicationCoord.setName(coordName);
             replicationCoord.setFrequency("${coord:" + feed.getFrequency().toString() + "}");
 
             long frequency_ms = ExpressionHelper.get().
                     evaluate(feed.getFrequency().toString(), Long.class);
             long timeout_ms = frequency_ms * 6;
-            if (timeout_ms < THIRTY_MINUTES) timeout_ms = THIRTY_MINUTES;
+            if (timeout_ms < THIRTY_MINUTES) {
+                timeout_ms = THIRTY_MINUTES;
+            }
             replicationCoord.getControls().setTimeout(String.valueOf(timeout_ms / (1000 * 60)));
             replicationCoord.getControls().setThrottle(String.valueOf(timeout_ms / frequency_ms * 2));
 
@@ -198,28 +199,30 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
             Date srcEndDate = FeedHelper.getCluster(feed, srcCluster.getName()).getValidity().getEnd();
             Date trgStartDate = FeedHelper.getCluster(feed, trgCluster.getName()).getValidity().getStart();
             Date trgEndDate = FeedHelper.getCluster(feed, trgCluster.getName()).getValidity().getEnd();
-			if (srcStartDate.after(trgEndDate)
-					|| trgStartDate.after(srcEndDate)) {
-				LOG.warn("Not creating replication coordinator, as the source cluster:"
-						+ srcCluster.getName()
-						+ " and target cluster: "
-						+ trgCluster.getName()
-						+ " do not have overlapping dates");
-				return null;
-			}
-            replicationCoord.setStart(srcStartDate.after(trgStartDate) ? SchemaHelper.formatDateUTC(srcStartDate) : SchemaHelper
-                    .formatDateUTC(trgStartDate));
-            replicationCoord.setEnd(srcEndDate.before(trgEndDate) ? SchemaHelper.formatDateUTC(srcEndDate) : SchemaHelper
-                    .formatDateUTC(trgEndDate));
+            if (srcStartDate.after(trgEndDate)
+                    || trgStartDate.after(srcEndDate)) {
+                LOG.warn("Not creating replication coordinator, as the source cluster:"
+                        + srcCluster.getName()
+                        + " and target cluster: "
+                        + trgCluster.getName()
+                        + " do not have overlapping dates");
+                return null;
+            }
+            replicationCoord.setStart(
+                    srcStartDate.after(trgStartDate) ? SchemaHelper.formatDateUTC(srcStartDate) : SchemaHelper
+                            .formatDateUTC(trgStartDate));
+            replicationCoord.setEnd(
+                    srcEndDate.before(trgEndDate) ? SchemaHelper.formatDateUTC(srcEndDate) : SchemaHelper
+                            .formatDateUTC(trgEndDate));
             replicationCoord.setTimezone(feed.getTimezone().getID());
             SYNCDATASET inputDataset = (SYNCDATASET) replicationCoord.getDatasets().getDatasetOrAsyncDataset().get(0);
             SYNCDATASET outputDataset = (SYNCDATASET) replicationCoord.getDatasets().getDatasetOrAsyncDataset().get(1);
 
-			inputDataset.setUriTemplate(new Path(ClusterHelper
-					.getStorageUrl(srcCluster), FeedHelper.getLocation(feed,
-					LocationType.DATA,srcCluster.getName()).getPath()).toString());
-			outputDataset.setUriTemplate(getStoragePath(FeedHelper.getLocation(
-					feed, LocationType.DATA, trgCluster.getName()).getPath()));
+            inputDataset.setUriTemplate(new Path(ClusterHelper
+                    .getStorageUrl(srcCluster), FeedHelper.getLocation(feed,
+                    LocationType.DATA, srcCluster.getName()).getPath()).toString());
+            outputDataset.setUriTemplate(getStoragePath(FeedHelper.getLocation(
+                    feed, LocationType.DATA, trgCluster.getName()).getPath()));
             setDatasetValues(inputDataset, feed, srcCluster);
             setDatasetValues(outputDataset, feed, srcCluster);
             if (feed.getAvailabilityFlag() == null) {
@@ -238,25 +241,30 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
     }
 
     private void setDatasetValues(SYNCDATASET dataset, Feed feed, Cluster cluster) {
-        dataset.setInitialInstance(SchemaHelper.formatDateUTC(FeedHelper.getCluster(feed, cluster.getName()).getValidity().getStart()));
+        dataset.setInitialInstance(
+                SchemaHelper.formatDateUTC(FeedHelper.getCluster(feed, cluster.getName()).getValidity().getStart()));
         dataset.setTimezone(feed.getTimezone().getID());
         dataset.setFrequency("${coord:" + feed.getFrequency().toString() + "}");
     }
 
-    private ACTION getReplicationWorkflowAction(Cluster srcCluster, Cluster trgCluster, Path wfPath, String wfName) throws FalconException {
+    private ACTION getReplicationWorkflowAction(Cluster srcCluster, Cluster trgCluster, Path wfPath, String wfName)
+            throws FalconException {
         ACTION replicationAction = new ACTION();
         WORKFLOW replicationWF = new WORKFLOW();
         try {
             replicationWF.setAppPath(getStoragePath(wfPath.toString()));
             Feed feed = getEntity();
 
-            String srcPart = FeedHelper.normalizePartitionExpression(FeedHelper.getCluster(feed, srcCluster.getName()).getPartition());
+            String srcPart = FeedHelper.normalizePartitionExpression(
+                    FeedHelper.getCluster(feed, srcCluster.getName()).getPartition());
             srcPart = FeedHelper.evaluateClusterExp(srcCluster, srcPart);
-            String targetPart = FeedHelper.normalizePartitionExpression(FeedHelper.getCluster(feed, trgCluster.getName()).getPartition());
+            String targetPart = FeedHelper.normalizePartitionExpression(
+                    FeedHelper.getCluster(feed, trgCluster.getName()).getPartition());
             targetPart = FeedHelper.evaluateClusterExp(trgCluster, targetPart);
-            
+
             StringBuilder pathsWithPartitions = new StringBuilder();
-            pathsWithPartitions.append("${coord:dataIn('input')}/").append(FeedHelper.normalizePartitionExpression(srcPart, targetPart));
+            pathsWithPartitions.append("${coord:dataIn('input')}/").append(
+                    FeedHelper.normalizePartitionExpression(srcPart, targetPart));
 
             Map<String, String> props = createCoordDefaultConfiguration(trgCluster, wfPath, wfName);
             props.put("srcClusterName", srcCluster.getName());
@@ -294,26 +302,27 @@ public class OozieFeedMapper extends AbstractOozieEntityMapper<Feed> {
         Feed feed = getEntity();
         Map<String, String> props = new HashMap<String, String>();
         if (feed.getProperties() != null) {
-            for (Property prop : feed.getProperties().getProperties())
+            for (Property prop : feed.getProperties().getProperties()) {
                 props.put(prop.getName(), prop.getValue());
+            }
         }
         return props;
     }
-    
-	private String getLocationURI(Cluster cluster, Feed feed, LocationType type) {
-		String path = FeedHelper.getLocation(feed, type, cluster.getName())
-				.getPath();
-
-		if (!path.equals("/tmp")) {
-			if (new Path(path).toUri().getScheme() == null){
-				return  new Path(ClusterHelper.getStorageUrl(cluster), path)
-						.toString();}
-			else{
-				return  path;
-			}
-		}
-		return null;
 
-}
+    private String getLocationURI(Cluster cluster, Feed feed, LocationType type) {
+        String path = FeedHelper.getLocation(feed, type, cluster.getName())
+                .getPath();
+
+        if (!path.equals("/tmp")) {
+            if (new Path(path).toUri().getScheme() == null) {
+                return new Path(ClusterHelper.getStorageUrl(cluster), path)
+                        .toString();
+            } else {
+                return path;
+            }
+        }
+        return null;
+
+    }
 
 }

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/feed/src/main/java/org/apache/falcon/workflow/OozieFeedWorkflowBuilder.java
----------------------------------------------------------------------
diff --git a/feed/src/main/java/org/apache/falcon/workflow/OozieFeedWorkflowBuilder.java b/feed/src/main/java/org/apache/falcon/workflow/OozieFeedWorkflowBuilder.java
index 3aa9eca..02cb740 100644
--- a/feed/src/main/java/org/apache/falcon/workflow/OozieFeedWorkflowBuilder.java
+++ b/feed/src/main/java/org/apache/falcon/workflow/OozieFeedWorkflowBuilder.java
@@ -18,13 +18,6 @@
 
 package org.apache.falcon.workflow;
 
-import java.util.Date;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.Properties;
-
-import org.apache.hadoop.fs.Path;
 import org.apache.falcon.FalconException;
 import org.apache.falcon.Tag;
 import org.apache.falcon.converter.AbstractOozieEntityMapper;
@@ -36,29 +29,37 @@ import org.apache.falcon.entity.v0.EntityType;
 import org.apache.falcon.entity.v0.cluster.Cluster;
 import org.apache.falcon.entity.v0.feed.Feed;
 import org.apache.falcon.security.CurrentUser;
+import org.apache.hadoop.fs.Path;
+
+import java.util.*;
 
 public class OozieFeedWorkflowBuilder extends OozieWorkflowBuilder<Feed> {
 
     @Override
     public Map<String, Properties> newWorkflowSchedule(Feed feed, List<String> clusters) throws FalconException {
         Map<String, Properties> propertiesMap = new HashMap<String, Properties>();
-        
-        for (String clusterName: clusters) {
+
+        for (String clusterName : clusters) {
             org.apache.falcon.entity.v0.feed.Cluster feedCluster = FeedHelper.getCluster(feed, clusterName);
-            Properties properties = newWorkflowSchedule(feed, feedCluster.getValidity().getStart(), clusterName, 
+            Properties properties = newWorkflowSchedule(feed, feedCluster.getValidity().getStart(), clusterName,
                     CurrentUser.getUser());
-            if (properties == null) continue;
+            if (properties == null) {
+                continue;
+            }
             propertiesMap.put(clusterName, properties);
         }
         return propertiesMap;
     }
 
     @Override
-    public Properties newWorkflowSchedule(Feed feed, Date startDate, String clusterName, String user) throws FalconException {
+    public Properties newWorkflowSchedule(Feed feed, Date startDate, String clusterName, String user)
+            throws FalconException {
         org.apache.falcon.entity.v0.feed.Cluster feedCluster = FeedHelper.getCluster(feed, clusterName);
         if (!startDate.before(feedCluster.getValidity().getEnd()))
-            // start time >= end time
+        // start time >= end time
+        {
             return null;
+        }
 
         Cluster cluster = configStore.get(EntityType.CLUSTER, feedCluster.getName());
         Path bundlePath = new Path(ClusterHelper.getLocation(cluster, "staging"), EntityUtil.getStagingPath(feed));
@@ -66,7 +67,7 @@ public class OozieFeedWorkflowBuilder extends OozieWorkflowBuilder<Feed> {
         EntityUtil.setStartDate(feedClone, clusterName, startDate);
 
         AbstractOozieEntityMapper<Feed> mapper = new OozieFeedMapper(feedClone);
-        if(!mapper.map(cluster, bundlePath)){
+        if (!mapper.map(cluster, bundlePath)) {
             return null;
         }
         return createAppProperties(clusterName, bundlePath, user);
@@ -79,10 +80,10 @@ public class OozieFeedWorkflowBuilder extends OozieWorkflowBuilder<Feed> {
                 feed.getFrequency(), feed.getTimezone(), now);
     }
 
-	@Override
-	public String[] getWorkflowNames(Feed entity) {
-		return new String[] {
-				EntityUtil.getWorkflowName(Tag.RETENTION, entity).toString(),
-				EntityUtil.getWorkflowName(Tag.REPLICATION, entity).toString() };
-	}
+    @Override
+    public String[] getWorkflowNames(Feed entity) {
+        return new String[]{
+                EntityUtil.getWorkflowName(Tag.RETENTION, entity).toString(),
+                EntityUtil.getWorkflowName(Tag.REPLICATION, entity).toString()};
+    }
 }

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/feed/src/main/resources/config/coordinator/replication-coordinator.xml
----------------------------------------------------------------------
diff --git a/feed/src/main/resources/config/coordinator/replication-coordinator.xml b/feed/src/main/resources/config/coordinator/replication-coordinator.xml
index ac1267a..693b0bd 100644
--- a/feed/src/main/resources/config/coordinator/replication-coordinator.xml
+++ b/feed/src/main/resources/config/coordinator/replication-coordinator.xml
@@ -16,36 +16,36 @@
   limitations under the License.
   -->
 <coordinator-app name="#NAME#" frequency="#FREQUENCY#"
-	start="#START_TIME#" end="#END_TIME" timezone="#TIMEZONE#"
-	xmlns="uri:oozie:coordinator:0.3">
-	<controls>
-		<concurrency>1</concurrency>
-		<execution>FIFO</execution>
-	</controls>
-	<datasets>
-		<dataset name="input-dataset" frequency="#FEED_FREQ#"
-			initial-instance="#START_TIME#" timezone="#TIMEZONE#">
-			<uri-template>#FEED_PATH#</uri-template>
-		</dataset>
-		<dataset name="output-dataset" frequency="#FEED_FREQ#"
-			initial-instance="#START_TIME#" timezone="#TIMEZONE#">
-			<uri-template>#FEED_PATH#</uri-template>
-		</dataset>
-	</datasets>
-	<input-events>
-		<data-in name="input" dataset="input-dataset">
-			<instance>${coord:current(0)}</instance>
-		</data-in>
-	</input-events>
-	<output-events>
-		<data-out name="output" dataset="output-dataset">
-			<instance>${coord:current(0)}</instance>
-		</data-out>
-	</output-events>
-	<action>
-		<workflow>
-			<app-path>#WF_PATH#</app-path>
-			<configuration />
-		</workflow>
-	</action>
+                 start="#START_TIME#" end="#END_TIME" timezone="#TIMEZONE#"
+                 xmlns="uri:oozie:coordinator:0.3">
+    <controls>
+        <concurrency>1</concurrency>
+        <execution>FIFO</execution>
+    </controls>
+    <datasets>
+        <dataset name="input-dataset" frequency="#FEED_FREQ#"
+                 initial-instance="#START_TIME#" timezone="#TIMEZONE#">
+            <uri-template>#FEED_PATH#</uri-template>
+        </dataset>
+        <dataset name="output-dataset" frequency="#FEED_FREQ#"
+                 initial-instance="#START_TIME#" timezone="#TIMEZONE#">
+            <uri-template>#FEED_PATH#</uri-template>
+        </dataset>
+    </datasets>
+    <input-events>
+        <data-in name="input" dataset="input-dataset">
+            <instance>${coord:current(0)}</instance>
+        </data-in>
+    </input-events>
+    <output-events>
+        <data-out name="output" dataset="output-dataset">
+            <instance>${coord:current(0)}</instance>
+        </data-out>
+    </output-events>
+    <action>
+        <workflow>
+            <app-path>#WF_PATH#</app-path>
+            <configuration/>
+        </workflow>
+    </action>
 </coordinator-app>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/feed/src/main/resources/config/workflow/replication-workflow.xml
----------------------------------------------------------------------
diff --git a/feed/src/main/resources/config/workflow/replication-workflow.xml b/feed/src/main/resources/config/workflow/replication-workflow.xml
index 5bb34e8..7aab158 100644
--- a/feed/src/main/resources/config/workflow/replication-workflow.xml
+++ b/feed/src/main/resources/config/workflow/replication-workflow.xml
@@ -16,99 +16,125 @@
   limitations under the License.
   -->
 <workflow-app xmlns='uri:oozie:workflow:0.3' name='falcon-feed-parent-workflow'>
-	<start to='should-record' />
-	<decision name='should-record'>
+    <start to='should-record'/>
+    <decision name='should-record'>
         <switch>
-            <case to="recordsize"> 
-              ${shouldRecord=="true"}
+            <case to="recordsize">
+                ${shouldRecord=="true"}
             </case>
             <default to="replication"/>
         </switch>
     </decision>
-	<action name='recordsize'>
-		<java>
-			<job-tracker>${jobTracker}</job-tracker>
-			<name-node>${nameNode}</name-node>
-			<configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
-			<main-class>org.apache.falcon.latedata.LateDataHandler</main-class>
-			<arg>-out</arg><arg>${logDir}/latedata/${nominalTime}/${srcClusterName}</arg>
-			<arg>-paths</arg><arg>${falconInPaths}</arg>
-			<arg>-falconInputFeeds</arg><arg>${falconInputFeeds}</arg>
-			<capture-output />
-		</java>
-		<ok to="replication" />
-		<error to="fail" />
-	</action>
+    <action name='recordsize'>
+        <java>
+            <job-tracker>${jobTracker}</job-tracker>
+            <name-node>${nameNode}</name-node>
+            <configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
+            <main-class>org.apache.falcon.latedata.LateDataHandler</main-class>
+            <arg>-out</arg>
+            <arg>${logDir}/latedata/${nominalTime}/${srcClusterName}</arg>
+            <arg>-paths</arg>
+            <arg>${falconInPaths}</arg>
+            <arg>-falconInputFeeds</arg>
+            <arg>${falconInputFeeds}</arg>
+            <capture-output/>
+        </java>
+        <ok to="replication"/>
+        <error to="fail"/>
+    </action>
     <action name="replication">
         <java>
             <job-tracker>${jobTracker}</job-tracker>
             <name-node>${nameNode}</name-node>
             <configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
             <main-class>org.apache.falcon.replication.FeedReplicator</main-class>
             <arg>-Dfalcon.include.path=${sourceRelativePaths}</arg>
             <arg>-Dmapred.job.queue.name=${queueName}</arg>
             <arg>-Dmapred.job.priority=${jobPriority}</arg>
-			<arg>-maxMaps</arg><arg>5</arg>
-			<arg>-sourcePaths</arg><arg>${distcpSourcePaths}</arg>
-			<arg>-targetPath</arg><arg>${distcpTargetPaths}</arg>
+            <arg>-maxMaps</arg>
+            <arg>5</arg>
+            <arg>-sourcePaths</arg>
+            <arg>${distcpSourcePaths}</arg>
+            <arg>-targetPath</arg>
+            <arg>${distcpTargetPaths}</arg>
             <file>${wf:conf("falcon.libpath")}/hadoop-distcp.jar</file>
         </java>
         <ok to="succeeded-post-processing"/>
         <error to="failed-post-processing"/>
     </action>
-   <action name='succeeded-post-processing'>
-		<java>
-			<job-tracker>${jobTracker}</job-tracker>
-			<name-node>${nameNode}</name-node>
-			<configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
-			<main-class>org.apache.falcon.workflow.FalconPostProcessing</main-class>
-			<arg>-cluster</arg><arg>${cluster}</arg>
-			<arg>-entityType</arg><arg>${entityType}</arg>
-			<arg>-entityName</arg><arg>${entityName}</arg>
-			<arg>-nominalTime</arg><arg>${nominalTime}</arg>
-			<arg>-operation</arg><arg>REPLICATE</arg>
-			<arg>-workflowId</arg><arg>${wf:id()}</arg>
-			<arg>-runId</arg><arg>${wf:run()}</arg>
-			<arg>-status</arg><arg>SUCCEEDED</arg>
-			<arg>-timeStamp</arg><arg>${timeStamp}</arg>			
-			<arg>-brokerImplClass</arg><arg>${wf:conf("broker.impl.class")}</arg>
-			<arg>-brokerUrl</arg><arg>${wf:conf("broker.url")}</arg>
-			<arg>-userBrokerImplClass</arg><arg>${userBrokerImplClass}</arg>
-			<arg>-userBrokerUrl</arg><arg>${userBrokerUrl}</arg>
-			<arg>-brokerTTL</arg><arg>${wf:conf("broker.ttlInMins")}</arg>
-			<arg>-feedNames</arg><arg>${feedNames}</arg>
-			<arg>-feedInstancePaths</arg><arg>${feedInstancePaths}</arg>			
-			<arg>-logFile</arg><arg>${logDir}/instancePaths-${nominalTime}-${srcClusterName}.csv</arg>
-			<arg>-workflowEngineUrl</arg> <arg>${workflowEngineUrl}</arg>
-			<arg>-subflowId</arg> <arg>${wf:id()}</arg>
-			<arg>-logDir</arg> <arg>${logDir}/job-${nominalTime}/${srcClusterName}/</arg>
+    <action name='succeeded-post-processing'>
+        <java>
+            <job-tracker>${jobTracker}</job-tracker>
+            <name-node>${nameNode}</name-node>
+            <configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
+            <main-class>org.apache.falcon.workflow.FalconPostProcessing</main-class>
+            <arg>-cluster</arg>
+            <arg>${cluster}</arg>
+            <arg>-entityType</arg>
+            <arg>${entityType}</arg>
+            <arg>-entityName</arg>
+            <arg>${entityName}</arg>
+            <arg>-nominalTime</arg>
+            <arg>${nominalTime}</arg>
+            <arg>-operation</arg>
+            <arg>REPLICATE</arg>
+            <arg>-workflowId</arg>
+            <arg>${wf:id()}</arg>
+            <arg>-runId</arg>
+            <arg>${wf:run()}</arg>
+            <arg>-status</arg>
+            <arg>SUCCEEDED</arg>
+            <arg>-timeStamp</arg>
+            <arg>${timeStamp}</arg>
+            <arg>-brokerImplClass</arg>
+            <arg>${wf:conf("broker.impl.class")}</arg>
+            <arg>-brokerUrl</arg>
+            <arg>${wf:conf("broker.url")}</arg>
+            <arg>-userBrokerImplClass</arg>
+            <arg>${userBrokerImplClass}</arg>
+            <arg>-userBrokerUrl</arg>
+            <arg>${userBrokerUrl}</arg>
+            <arg>-brokerTTL</arg>
+            <arg>${wf:conf("broker.ttlInMins")}</arg>
+            <arg>-feedNames</arg>
+            <arg>${feedNames}</arg>
+            <arg>-feedInstancePaths</arg>
+            <arg>${feedInstancePaths}</arg>
+            <arg>-logFile</arg>
+            <arg>${logDir}/instancePaths-${nominalTime}-${srcClusterName}.csv</arg>
+            <arg>-workflowEngineUrl</arg>
+            <arg>${workflowEngineUrl}</arg>
+            <arg>-subflowId</arg>
+            <arg>${wf:id()}</arg>
+            <arg>-logDir</arg>
+            <arg>${logDir}/job-${nominalTime}/${srcClusterName}/</arg>
             <file>${wf:conf("falcon.libpath")}/activemq-core.jar</file>
             <file>${wf:conf("falcon.libpath")}/ant.jar</file>
             <file>${wf:conf("falcon.libpath")}/geronimo-j2ee-management.jar</file>
@@ -116,45 +142,65 @@
             <file>${wf:conf("falcon.libpath")}/json-simple.jar</file>
             <file>${wf:conf("falcon.libpath")}/oozie-client.jar</file>
             <file>${wf:conf("falcon.libpath")}/spring-jms.jar</file>
-		</java>
-		<ok to="end" />
-		<error to="fail" />
-	</action>
-		<action name='failed-post-processing'>
-		<java>
-			<job-tracker>${jobTracker}</job-tracker>
-			<name-node>${nameNode}</name-node>
-			<configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
-			<main-class>org.apache.falcon.workflow.FalconPostProcessing</main-class>
-			<arg>-cluster</arg><arg>${cluster}</arg>
-			<arg>-entityType</arg><arg>${entityType}</arg>
-			<arg>-entityName</arg><arg>${entityName}</arg>
-			<arg>-nominalTime</arg><arg>${nominalTime}</arg>
-			<arg>-operation</arg><arg>REPLICATE</arg>
-			<arg>-workflowId</arg><arg>${wf:id()}</arg>
-			<arg>-runId</arg><arg>${wf:run()}</arg>
-			<arg>-status</arg><arg>FAILED</arg>
-			<arg>-timeStamp</arg><arg>${timeStamp}</arg>			
-			<arg>-brokerImplClass</arg><arg>${wf:conf("broker.impl.class")}</arg>
-			<arg>-brokerUrl</arg><arg>${wf:conf("broker.url")}</arg>
-			<arg>-userBrokerImplClass</arg><arg>${userBrokerImplClass}</arg>
-			<arg>-userBrokerUrl</arg><arg>${userBrokerUrl}</arg>
-			<arg>-brokerTTL</arg><arg>${wf:conf("broker.ttlInMins")}</arg>
-			<arg>-feedNames</arg><arg>${feedNames}</arg>
-			<arg>-feedInstancePaths</arg><arg>${feedInstancePaths}</arg>			
-			<arg>-logFile</arg><arg>${logDir}/instancePaths-${nominalTime}-${srcClusterName}.csv</arg>
-			<arg>-workflowEngineUrl</arg> <arg>${workflowEngineUrl}</arg>
-			<arg>-subflowId</arg> <arg>${wf:id()}</arg>
-			<arg>-logDir</arg> <arg>${logDir}/job-${nominalTime}/${srcClusterName}/</arg>
+        </java>
+        <ok to="end"/>
+        <error to="fail"/>
+    </action>
+    <action name='failed-post-processing'>
+        <java>
+            <job-tracker>${jobTracker}</job-tracker>
+            <name-node>${nameNode}</name-node>
+            <configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
+            <main-class>org.apache.falcon.workflow.FalconPostProcessing</main-class>
+            <arg>-cluster</arg>
+            <arg>${cluster}</arg>
+            <arg>-entityType</arg>
+            <arg>${entityType}</arg>
+            <arg>-entityName</arg>
+            <arg>${entityName}</arg>
+            <arg>-nominalTime</arg>
+            <arg>${nominalTime}</arg>
+            <arg>-operation</arg>
+            <arg>REPLICATE</arg>
+            <arg>-workflowId</arg>
+            <arg>${wf:id()}</arg>
+            <arg>-runId</arg>
+            <arg>${wf:run()}</arg>
+            <arg>-status</arg>
+            <arg>FAILED</arg>
+            <arg>-timeStamp</arg>
+            <arg>${timeStamp}</arg>
+            <arg>-brokerImplClass</arg>
+            <arg>${wf:conf("broker.impl.class")}</arg>
+            <arg>-brokerUrl</arg>
+            <arg>${wf:conf("broker.url")}</arg>
+            <arg>-userBrokerImplClass</arg>
+            <arg>${userBrokerImplClass}</arg>
+            <arg>-userBrokerUrl</arg>
+            <arg>${userBrokerUrl}</arg>
+            <arg>-brokerTTL</arg>
+            <arg>${wf:conf("broker.ttlInMins")}</arg>
+            <arg>-feedNames</arg>
+            <arg>${feedNames}</arg>
+            <arg>-feedInstancePaths</arg>
+            <arg>${feedInstancePaths}</arg>
+            <arg>-logFile</arg>
+            <arg>${logDir}/instancePaths-${nominalTime}-${srcClusterName}.csv</arg>
+            <arg>-workflowEngineUrl</arg>
+            <arg>${workflowEngineUrl}</arg>
+            <arg>-subflowId</arg>
+            <arg>${wf:id()}</arg>
+            <arg>-logDir</arg>
+            <arg>${logDir}/job-${nominalTime}/${srcClusterName}/</arg>
             <file>${wf:conf("falcon.libpath")}/activemq-core.jar</file>
             <file>${wf:conf("falcon.libpath")}/ant.jar</file>
             <file>${wf:conf("falcon.libpath")}/geronimo-j2ee-management.jar</file>
@@ -162,14 +208,14 @@
             <file>${wf:conf("falcon.libpath")}/json-simple.jar</file>
             <file>${wf:conf("falcon.libpath")}/oozie-client.jar</file>
             <file>${wf:conf("falcon.libpath")}/spring-jms.jar</file>
-		</java>
-		<ok to="fail" />
-		<error to="fail" />
-	</action>
-	<kill name="fail">
-		<message>Workflow failed, error
-			message[${wf:errorMessage(wf:lastErrorNode())}]
-		</message>
-	</kill>
-	<end name='end' />
+        </java>
+        <ok to="fail"/>
+        <error to="fail"/>
+    </action>
+    <kill name="fail">
+        <message>Workflow failed, error
+            message[${wf:errorMessage(wf:lastErrorNode())}]
+        </message>
+    </kill>
+    <end name='end'/>
 </workflow-app>

http://git-wip-us.apache.org/repos/asf/incubator-falcon/blob/a4d79f0c/feed/src/main/resources/config/workflow/retention-workflow.xml
----------------------------------------------------------------------
diff --git a/feed/src/main/resources/config/workflow/retention-workflow.xml b/feed/src/main/resources/config/workflow/retention-workflow.xml
index 73f5d36..422209e 100644
--- a/feed/src/main/resources/config/workflow/retention-workflow.xml
+++ b/feed/src/main/resources/config/workflow/retention-workflow.xml
@@ -16,64 +16,86 @@
   limitations under the License.
   -->
 <workflow-app xmlns='uri:oozie:workflow:0.3' name='falcon-feed-parent-workflow'>
-	<start to='eviction' />
+    <start to='eviction'/>
     <action name="eviction">
         <java>
             <job-tracker>${jobTracker}</job-tracker>
             <name-node>${nameNode}</name-node>
             <configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
             <main-class>org.apache.falcon.retention.FeedEvictor</main-class>
-            <arg>-feedBasePath</arg><arg>${feedDataPath}</arg>
-            <arg>-retentionType</arg><arg>instance</arg>
-            <arg>-retentionLimit</arg><arg>${limit}</arg>
-            <arg>-timeZone</arg><arg>${timeZone}</arg>
-            <arg>-frequency</arg><arg>${frequency}</arg>
-			<arg>-logFile</arg><arg>${logDir}/instancePaths-${nominalTime}.csv</arg>
+            <arg>-feedBasePath</arg>
+            <arg>${feedDataPath}</arg>
+            <arg>-retentionType</arg>
+            <arg>instance</arg>
+            <arg>-retentionLimit</arg>
+            <arg>${limit}</arg>
+            <arg>-timeZone</arg>
+            <arg>${timeZone}</arg>
+            <arg>-frequency</arg>
+            <arg>${frequency}</arg>
+            <arg>-logFile</arg>
+            <arg>${logDir}/instancePaths-${nominalTime}.csv</arg>
         </java>
         <ok to="jms-messaging"/>
         <error to="fail"/>
     </action>
-	
-	<action name='jms-messaging'>
-		<java>
-			<job-tracker>${jobTracker}</job-tracker>
-			<name-node>${nameNode}</name-node>
-			<configuration>
-				<property>
-					<name>mapred.job.queue.name</name>
-					<value>${queueName}</value>
-				</property>
-				<property>
-					<name>oozie.launcher.mapred.job.priority</name>
-					<value>${jobPriority}</value>
-				</property>
-			</configuration>
-			<main-class>org.apache.falcon.messaging.MessageProducer</main-class>
-			<arg>-entityName</arg><arg>${entityName}</arg>
-			<arg>-feedNames</arg><arg>${feedNames}</arg>
-			<arg>-feedInstancePaths</arg><arg>${feedInstancePaths}</arg>
-			<arg>-workflowId</arg><arg>${wf:id()}</arg>
-			<arg>-runId</arg><arg>${wf:run()}</arg>
-			<arg>-nominalTime</arg><arg>${nominalTime}</arg>
-			<arg>-timeStamp</arg><arg>${timeStamp}</arg>
-			<arg>-brokerUrl</arg><arg>${userBrokerUrl}</arg>
-			<arg>-brokerImplClass</arg><arg>${userBrokerImplClass}</arg>
-			<arg>-entityType</arg><arg>${entityType}</arg>
-			<arg>-operation</arg><arg>DELETE</arg>
-			<arg>-logFile</arg><arg>${logDir}/instancePaths-${nominalTime}.csv</arg>
-			<arg>-topicName</arg><arg>FALCON.${entityName}</arg>
-			<arg>-status</arg><arg>SUCCEEDED</arg>
-			<arg>-brokerTTL</arg><arg>${wf:conf("broker.ttlInMins")}</arg>
-			<arg>-cluster</arg><arg>${cluster}</arg>
+
+    <action name='jms-messaging'>
+        <java>
+            <job-tracker>${jobTracker}</job-tracker>
+            <name-node>${nameNode}</name-node>
+            <configuration>
+                <property>
+                    <name>mapred.job.queue.name</name>
+                    <value>${queueName}</value>
+                </property>
+                <property>
+                    <name>oozie.launcher.mapred.job.priority</name>
+                    <value>${jobPriority}</value>
+                </property>
+            </configuration>
+            <main-class>org.apache.falcon.messaging.MessageProducer</main-class>
+            <arg>-entityName</arg>
+            <arg>${entityName}</arg>
+            <arg>-feedNames</arg>
+            <arg>${feedNames}</arg>
+            <arg>-feedInstancePaths</arg>
+            <arg>${feedInstancePaths}</arg>
+            <arg>-workflowId</arg>
+            <arg>${wf:id()}</arg>
+            <arg>-runId</arg>
+            <arg>${wf:run()}</arg>
+            <arg>-nominalTime</arg>
+            <arg>${nominalTime}</arg>
+            <arg>-timeStamp</arg>
+            <arg>${timeStamp}</arg>
+            <arg>-brokerUrl</arg>
+            <arg>${userBrokerUrl}</arg>
+            <arg>-brokerImplClass</arg>
+            <arg>${userBrokerImplClass}</arg>
+            <arg>-entityType</arg>
+            <arg>${entityType}</arg>
+            <arg>-operation</arg>
+            <arg>DELETE</arg>
+            <arg>-logFile</arg>
+            <arg>${logDir}/instancePaths-${nominalTime}.csv</arg>
+            <arg>-topicName</arg>
+            <arg>FALCON.${entityName}</arg>
+            <arg>-status</arg>
+            <arg>SUCCEEDED</arg>
+            <arg>-brokerTTL</arg>
+            <arg>${wf:conf("broker.ttlInMins")}</arg>
+            <arg>-cluster</arg>
+            <arg>${cluster}</arg>
             <file>${wf:conf("falcon.libpath")}/activemq-core.jar</file>
             <file>${wf:conf("falcon.libpath")}/ant.jar</file>
             <file>${wf:conf("falcon.libpath")}/geronimo-j2ee-management.jar</file>
@@ -81,15 +103,15 @@
             <file>${wf:conf("falcon.libpath")}/json-simple.jar</file>
             <file>${wf:conf("falcon.libpath")}/oozie-client.jar</file>
             <file>${wf:conf("falcon.libpath")}/spring-jms.jar</file>
-		</java>
-		<ok to="end" />
-		<error to="fail" />
-	</action>	
+        </java>
+        <ok to="end"/>
+        <error to="fail"/>
+    </action>
 
-	<kill name="fail">
-		<message>Workflow failed, error
-			message[${wf:errorMessage(wf:lastErrorNode())}]
-		</message>
-	</kill>
-	<end name='end' />
+    <kill name="fail">
+        <message>Workflow failed, error
+            message[${wf:errorMessage(wf:lastErrorNode())}]
+        </message>
+    </kill>
+    <end name='end'/>
 </workflow-app>


Mime
View raw message