accumulo-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From dlmar...@apache.org
Subject [3/3] accumulo git commit: Merge branch '1.7' into 1.8
Date Fri, 20 May 2016 17:10:50 GMT
Merge branch '1.7' into 1.8


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/135bf025
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/135bf025
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/135bf025

Branch: refs/heads/1.8
Commit: 135bf0250752f76a32bb48703676e928ccf42452
Parents: d950d35 b3df354
Author: Dave Marion <dlmarion@apache.org>
Authored: Fri May 20 12:48:39 2016 -0400
Committer: Dave Marion <dlmarion@apache.org>
Committed: Fri May 20 12:48:39 2016 -0400

----------------------------------------------------------------------
 pom.xml                                         |   2 +-
 .../classloader/vfs/AccumuloVFSClassLoader.java |   4 +-
 .../vfs/providers/HdfsFileAttributes.java       |  59 -----
 .../providers/HdfsFileContentInfoFactory.java   |  50 ----
 .../vfs/providers/HdfsFileObject.java           | 240 -------------------
 .../vfs/providers/HdfsFileProvider.java         |  65 -----
 .../vfs/providers/HdfsFileSystem.java           | 126 ----------
 .../providers/HdfsFileSystemConfigBuilder.java  |  45 ----
 .../vfs/providers/HdfsRandomAccessContent.java  | 228 ------------------
 .../classloader/vfs/providers/package.html      |  19 --
 .../providers/ReadOnlyHdfsFileProviderTest.java |   2 +
 .../accumulo/start/test/AccumuloDFSBase.java    |   2 +-
 12 files changed, 6 insertions(+), 836 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/135bf025/pom.xml
----------------------------------------------------------------------
diff --cc pom.xml
index e983f11,d393fcd..8241b64
--- a/pom.xml
+++ b/pom.xml
@@@ -376,19 -356,9 +376,19 @@@
        <dependency>
          <groupId>org.apache.commons</groupId>
          <artifactId>commons-vfs2</artifactId>
-         <version>2.0</version>
+         <version>2.1</version>
        </dependency>
        <dependency>
 +        <groupId>org.apache.curator</groupId>
 +        <artifactId>curator-framework</artifactId>
 +        <version>${curator.version}</version>
 +      </dependency>
 +      <dependency>
 +        <groupId>org.apache.curator</groupId>
 +        <artifactId>curator-test</artifactId>
 +        <version>${curator.version}</version>
 +      </dependency>
 +      <dependency>
          <groupId>org.apache.hadoop</groupId>
          <artifactId>hadoop-client</artifactId>
          <version>${hadoop.version}</version>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/135bf025/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/135bf025/start/src/test/java/org/apache/accumulo/start/test/AccumuloDFSBase.java
----------------------------------------------------------------------
diff --cc start/src/test/java/org/apache/accumulo/start/test/AccumuloDFSBase.java
index fc5c16d,0000000..b405466
mode 100644,000000..100644
--- a/start/src/test/java/org/apache/accumulo/start/test/AccumuloDFSBase.java
+++ b/start/src/test/java/org/apache/accumulo/start/test/AccumuloDFSBase.java
@@@ -1,130 -1,0 +1,130 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.start.test;
 +
 +import java.io.File;
 +import java.io.IOException;
 +import java.net.URI;
 +
 +import org.apache.accumulo.start.classloader.vfs.MiniDFSUtil;
- import org.apache.accumulo.start.classloader.vfs.providers.HdfsFileProvider;
 +import org.apache.commons.vfs2.CacheStrategy;
 +import org.apache.commons.vfs2.FileSystemException;
 +import org.apache.commons.vfs2.cache.DefaultFilesCache;
 +import org.apache.commons.vfs2.cache.SoftRefFilesCache;
 +import org.apache.commons.vfs2.impl.DefaultFileReplicator;
 +import org.apache.commons.vfs2.impl.DefaultFileSystemManager;
 +import org.apache.commons.vfs2.impl.FileContentInfoFilenameFactory;
++import org.apache.commons.vfs2.provider.hdfs.HdfsFileProvider;
 +import org.apache.hadoop.conf.Configuration;
 +import org.apache.hadoop.hdfs.DFSConfigKeys;
 +import org.apache.hadoop.hdfs.MiniDFSCluster;
 +import org.junit.AfterClass;
 +import org.junit.BeforeClass;
 +
 +public class AccumuloDFSBase {
 +
 +  protected static Configuration conf = null;
 +  protected static DefaultFileSystemManager vfs = null;
 +  protected static MiniDFSCluster cluster = null;
 +
 +  private static URI HDFS_URI;
 +
 +  protected static URI getHdfsUri() {
 +    return HDFS_URI;
 +  }
 +
 +  @BeforeClass
 +  public static void miniDfsClusterSetup() {
 +    System.setProperty("java.io.tmpdir", System.getProperty("user.dir") + "/target");
 +    // System.setProperty("org.apache.commons.logging.Log", "org.apache.commons.logging.impl.NoOpLog");
 +    // Logger.getRootLogger().setLevel(Level.ERROR);
 +
 +    // Put the MiniDFSCluster directory in the target directory
 +    System.setProperty("test.build.data", "target/build/test/data");
 +
 +    // Setup HDFS
 +    conf = new Configuration();
 +    conf.set("hadoop.security.token.service.use_ip", "true");
 +
 +    conf.set("dfs.datanode.data.dir.perm", MiniDFSUtil.computeDatanodeDirectoryPermission());
 +    conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 1024 * 1024); // 1M blocksize
 +
 +    try {
 +      cluster = new MiniDFSCluster.Builder(conf).build();
 +      cluster.waitClusterUp();
 +      // We can't assume that the hostname of "localhost" will still be "localhost" after
 +      // starting up the NameNode. We may get mapped into a FQDN via settings in /etc/hosts.
 +      HDFS_URI = cluster.getFileSystem().getUri();
 +    } catch (IOException e) {
 +      throw new RuntimeException("Error setting up mini cluster", e);
 +    }
 +
 +    // Set up the VFS
 +    vfs = new DefaultFileSystemManager();
 +    try {
 +      vfs.setFilesCache(new DefaultFilesCache());
 +      vfs.addProvider("res", new org.apache.commons.vfs2.provider.res.ResourceFileProvider());
 +      vfs.addProvider("zip", new org.apache.commons.vfs2.provider.zip.ZipFileProvider());
 +      vfs.addProvider("gz", new org.apache.commons.vfs2.provider.gzip.GzipFileProvider());
 +      vfs.addProvider("ram", new org.apache.commons.vfs2.provider.ram.RamFileProvider());
 +      vfs.addProvider("file", new org.apache.commons.vfs2.provider.local.DefaultLocalFileProvider());
 +      vfs.addProvider("jar", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("http", new org.apache.commons.vfs2.provider.http.HttpFileProvider());
 +      vfs.addProvider("https", new org.apache.commons.vfs2.provider.https.HttpsFileProvider());
 +      vfs.addProvider("ftp", new org.apache.commons.vfs2.provider.ftp.FtpFileProvider());
 +      vfs.addProvider("ftps", new org.apache.commons.vfs2.provider.ftps.FtpsFileProvider());
 +      vfs.addProvider("war", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("par", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("ear", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("sar", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("ejb3", new org.apache.commons.vfs2.provider.jar.JarFileProvider());
 +      vfs.addProvider("tmp", new org.apache.commons.vfs2.provider.temp.TemporaryFileProvider());
 +      vfs.addProvider("tar", new org.apache.commons.vfs2.provider.tar.TarFileProvider());
 +      vfs.addProvider("tbz2", new org.apache.commons.vfs2.provider.tar.TarFileProvider());
 +      vfs.addProvider("tgz", new org.apache.commons.vfs2.provider.tar.TarFileProvider());
 +      vfs.addProvider("bz2", new org.apache.commons.vfs2.provider.bzip2.Bzip2FileProvider());
 +      vfs.addProvider("hdfs", new HdfsFileProvider());
 +      vfs.addExtensionMap("jar", "jar");
 +      vfs.addExtensionMap("zip", "zip");
 +      vfs.addExtensionMap("gz", "gz");
 +      vfs.addExtensionMap("tar", "tar");
 +      vfs.addExtensionMap("tbz2", "tar");
 +      vfs.addExtensionMap("tgz", "tar");
 +      vfs.addExtensionMap("bz2", "bz2");
 +      vfs.addMimeTypeMap("application/x-tar", "tar");
 +      vfs.addMimeTypeMap("application/x-gzip", "gz");
 +      vfs.addMimeTypeMap("application/zip", "zip");
 +      vfs.setFileContentInfoFactory(new FileContentInfoFilenameFactory());
 +      vfs.setFilesCache(new SoftRefFilesCache());
 +      vfs.setReplicator(new DefaultFileReplicator(new File(System.getProperty("java.io.tmpdir"),
"accumulo-vfs-cache-"
 +          + System.getProperty("user.name", "nouser"))));
 +      vfs.setCacheStrategy(CacheStrategy.ON_RESOLVE);
 +      vfs.init();
 +    } catch (FileSystemException e) {
 +      throw new RuntimeException("Error setting up VFS", e);
 +    }
 +
 +  }
 +
 +  @AfterClass
 +  public static void tearDownMiniDfsCluster() {
 +    if (null != cluster) {
 +      cluster.shutdown();
 +    }
 +  }
 +
 +}


Mime
View raw message