Return-Path: X-Original-To: apmail-hadoop-mapreduce-commits-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-commits-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id BB7E9DD88 for ; Mon, 13 Aug 2012 20:05:35 +0000 (UTC) Received: (qmail 95380 invoked by uid 500); 13 Aug 2012 20:05:35 -0000 Delivered-To: apmail-hadoop-mapreduce-commits-archive@hadoop.apache.org Received: (qmail 95295 invoked by uid 500); 13 Aug 2012 20:05:35 -0000 Mailing-List: contact mapreduce-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: mapreduce-dev@hadoop.apache.org Delivered-To: mailing list mapreduce-commits@hadoop.apache.org Received: (qmail 95279 invoked by uid 99); 13 Aug 2012 20:05:34 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 13 Aug 2012 20:05:34 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 13 Aug 2012 20:05:33 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id C627723888E4; Mon, 13 Aug 2012 20:04:49 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1372573 - in /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project: ./ hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/ hadoop-mapreduce-client/hadoop-mapreduce-client-common/... Date: Mon, 13 Aug 2012 20:04:49 -0000 To: mapreduce-commits@hadoop.apache.org From: bobby@apache.org X-Mailer: svnmailer-1.0.8-patched Message-Id: <20120813200449.C627723888E4@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Author: bobby Date: Mon Aug 13 20:04:49 2012 New Revision: 1372573 URL: http://svn.apache.org/viewvc?rev=1372573&view=rev Log: svn merge -c -1369197 Reverting: MAPREDUCE-4503 in branch-0.23 until MAPREDUCE-4549 can be addressed. Modified: hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/v2/util/TestMRApps.java Modified: hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt?rev=1372573&r1=1372572&r2=1372573&view=diff ============================================================================== --- hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt (original) +++ hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt Mon Aug 13 20:04:49 2012 @@ -382,9 +382,6 @@ Release 0.23.3 - UNRELEASED MAPREDUCE-4504. SortValidator writes to wrong directory (Robert Evans via tgraves) - MAPREDUCE-4503. Should throw InvalidJobConfException if duplicates found in - cacheArchives or cacheFiles (Robert Evans via jeagles) - MAPREDUCE-3782. teragen terasort jobs fail when using webhdfs:// (Jason Lowe via bobby) Modified: hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java?rev=1372573&r1=1372572&r2=1372573&view=diff ============================================================================== --- hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java (original) +++ hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java Mon Aug 13 20:04:49 2012 @@ -35,7 +35,6 @@ import org.apache.hadoop.classification. import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.mapred.InvalidJobConfException; import org.apache.hadoop.mapreduce.JobID; import org.apache.hadoop.mapreduce.MRJobConfig; import org.apache.hadoop.mapreduce.TaskAttemptID; @@ -57,7 +56,6 @@ import org.apache.hadoop.yarn.api.record import org.apache.hadoop.yarn.conf.YarnConfiguration; import org.apache.hadoop.yarn.util.Apps; import org.apache.hadoop.yarn.util.BuilderUtils; -import org.apache.hadoop.yarn.util.ConverterUtils; /** * Helper class for MR applications @@ -265,13 +263,6 @@ public class MRApps extends Apps { DistributedCache.getFileClassPaths(conf)); } - private static String getResourceDescription(LocalResourceType type) { - if(type == LocalResourceType.ARCHIVE) { - return "cache archive (" + MRJobConfig.CACHE_ARCHIVES + ") "; - } - return "cache file (" + MRJobConfig.CACHE_FILES + ") "; - } - // TODO - Move this to MR! // Use TaskDistributedCacheManager.CacheFiles.makeCacheFiles(URI[], // long[], boolean[], Path[], FileType) @@ -317,13 +308,6 @@ public class MRApps extends Apps { throw new IllegalArgumentException("Resource name must be relative"); } String linkName = name.toUri().getPath(); - LocalResource orig = localResources.get(linkName); - if(orig != null && !orig.getResource().equals( - ConverterUtils.getYarnUrlFromURI(p.toUri()))) { - throw new InvalidJobConfException( - getResourceDescription(orig.getType()) + orig.getResource() + - " conflicts with " + getResourceDescription(type) + u); - } localResources.put( linkName, BuilderUtils.newLocalResource( Modified: hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/v2/util/TestMRApps.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/v2/util/TestMRApps.java?rev=1372573&r1=1372572&r2=1372573&view=diff ============================================================================== --- hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/v2/util/TestMRApps.java (original) +++ hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/v2/util/TestMRApps.java Mon Aug 13 20:04:49 2012 @@ -19,19 +19,13 @@ package org.apache.hadoop.mapreduce.v2.util; import java.io.IOException; -import java.net.URI; import java.util.HashMap; import java.util.Map; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.FilterFileSystem; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.mapred.InvalidJobConfException; import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.JobID; import org.apache.hadoop.mapreduce.MRJobConfig; -import org.apache.hadoop.mapreduce.filecache.DistributedCache; import org.apache.hadoop.mapreduce.v2.api.records.JobId; import org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptId; import org.apache.hadoop.mapreduce.v2.api.records.TaskId; @@ -39,12 +33,9 @@ import org.apache.hadoop.mapreduce.v2.ap import org.apache.hadoop.mapreduce.v2.util.MRApps; import org.apache.hadoop.yarn.api.records.ApplicationId; import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider; -import org.apache.hadoop.yarn.api.records.LocalResource; -import org.apache.hadoop.yarn.api.records.LocalResourceType; import org.junit.Test; import static org.junit.Assert.*; -import static org.mockito.Mockito.*; public class TestMRApps { @@ -177,122 +168,5 @@ public class TestMRApps { assertNotSame("MAPREDUCE_JOB_USER_CLASSPATH_FIRST false, but taking effect!", env_str.indexOf("$PWD:job.jar"), 0); } - - @Test - public void testSetupDistributedCacheEmpty() throws IOException { - Configuration conf = new Configuration(); - Map localResources = new HashMap(); - MRApps.setupDistributedCache(conf, localResources); - assertTrue("Empty Config did not produce an empty list of resources", - localResources.isEmpty()); - } - - @SuppressWarnings("deprecation") - @Test(expected = InvalidJobConfException.class) - public void testSetupDistributedCacheConflicts() throws Exception { - Configuration conf = new Configuration(); - conf.setClass("fs.mockfs.impl", MockFileSystem.class, FileSystem.class); - - URI mockUri = URI.create("mockfs://mock/"); - FileSystem mockFs = ((FilterFileSystem)FileSystem.get(mockUri, conf)) - .getRawFileSystem(); - - URI archive = new URI("mockfs://mock/tmp/something.zip#something"); - Path archivePath = new Path(archive); - URI file = new URI("mockfs://mock/tmp/something.txt#something"); - Path filePath = new Path(file); - - when(mockFs.resolvePath(archivePath)).thenReturn(archivePath); - when(mockFs.resolvePath(filePath)).thenReturn(filePath); - - DistributedCache.addCacheArchive(archive, conf); - conf.set(MRJobConfig.CACHE_ARCHIVES_TIMESTAMPS, "10"); - conf.set(MRJobConfig.CACHE_ARCHIVES_SIZES, "10"); - conf.set(MRJobConfig.CACHE_ARCHIVES_VISIBILITIES, "true"); - DistributedCache.addCacheFile(file, conf); - conf.set(MRJobConfig.CACHE_FILE_TIMESTAMPS, "11"); - conf.set(MRJobConfig.CACHE_FILES_SIZES, "11"); - conf.set(MRJobConfig.CACHE_FILE_VISIBILITIES, "true"); - Map localResources = - new HashMap(); - MRApps.setupDistributedCache(conf, localResources); - } - - @SuppressWarnings("deprecation") - @Test(expected = InvalidJobConfException.class) - public void testSetupDistributedCacheConflictsFiles() throws Exception { - Configuration conf = new Configuration(); - conf.setClass("fs.mockfs.impl", MockFileSystem.class, FileSystem.class); - - URI mockUri = URI.create("mockfs://mock/"); - FileSystem mockFs = ((FilterFileSystem)FileSystem.get(mockUri, conf)) - .getRawFileSystem(); - - URI file = new URI("mockfs://mock/tmp/something.zip#something"); - Path filePath = new Path(file); - URI file2 = new URI("mockfs://mock/tmp/something.txt#something"); - Path file2Path = new Path(file); - - when(mockFs.resolvePath(filePath)).thenReturn(filePath); - when(mockFs.resolvePath(file2Path)).thenReturn(file2Path); - - DistributedCache.addCacheFile(file, conf); - DistributedCache.addCacheFile(file2, conf); - conf.set(MRJobConfig.CACHE_FILE_TIMESTAMPS, "10,11"); - conf.set(MRJobConfig.CACHE_FILES_SIZES, "10,11"); - conf.set(MRJobConfig.CACHE_FILE_VISIBILITIES, "true,true"); - Map localResources = - new HashMap(); - MRApps.setupDistributedCache(conf, localResources); - } - - @SuppressWarnings("deprecation") - @Test - public void testSetupDistributedCache() throws Exception { - Configuration conf = new Configuration(); - conf.setClass("fs.mockfs.impl", MockFileSystem.class, FileSystem.class); - - URI mockUri = URI.create("mockfs://mock/"); - FileSystem mockFs = ((FilterFileSystem)FileSystem.get(mockUri, conf)) - .getRawFileSystem(); - - URI archive = new URI("mockfs://mock/tmp/something.zip"); - Path archivePath = new Path(archive); - URI file = new URI("mockfs://mock/tmp/something.txt#something"); - Path filePath = new Path(file); - - when(mockFs.resolvePath(archivePath)).thenReturn(archivePath); - when(mockFs.resolvePath(filePath)).thenReturn(filePath); - - DistributedCache.addCacheArchive(archive, conf); - conf.set(MRJobConfig.CACHE_ARCHIVES_TIMESTAMPS, "10"); - conf.set(MRJobConfig.CACHE_ARCHIVES_SIZES, "10"); - conf.set(MRJobConfig.CACHE_ARCHIVES_VISIBILITIES, "true"); - DistributedCache.addCacheFile(file, conf); - conf.set(MRJobConfig.CACHE_FILE_TIMESTAMPS, "11"); - conf.set(MRJobConfig.CACHE_FILES_SIZES, "11"); - conf.set(MRJobConfig.CACHE_FILE_VISIBILITIES, "true"); - Map localResources = - new HashMap(); - MRApps.setupDistributedCache(conf, localResources); - assertEquals(2, localResources.size()); - LocalResource lr = localResources.get("something.zip"); - assertNotNull(lr); - assertEquals(10l, lr.getSize()); - assertEquals(10l, lr.getTimestamp()); - assertEquals(LocalResourceType.ARCHIVE, lr.getType()); - lr = localResources.get("something"); - assertNotNull(lr); - assertEquals(11l, lr.getSize()); - assertEquals(11l, lr.getTimestamp()); - assertEquals(LocalResourceType.FILE, lr.getType()); - } - - static class MockFileSystem extends FilterFileSystem { - MockFileSystem() { - super(mock(FileSystem.class)); - } - public void initialize(URI name, Configuration conf) throws IOException {} - } - + }