mahout-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sean Owen <sro...@gmail.com>
Subject Re: StackOverflowError on running bin/mahout with HADOOP_CONF_DIR specified
Date Sat, 02 Jul 2011 22:04:04 GMT
Looks like a Hadoop bug as far as I can tell... or even something
weird about your local config or file system. It's failing somehow
while reading an XML file from the file system?

It's not from Mahout in any event.

On Sat, Jul 2, 2011 at 8:01 PM, Sergey Bartunov <sbos.net@gmail.com> wrote:
> All the day I'm trying to do something around the
> http://stackoverflow.com/questions/6558606/hadoop-0-20-203-dont-load-configuration-files
>
> And now I pass the HADOOP_CONF_DIR to mahout to ensure that I include
> all necessary configuration files. But if do this, mahout fails with
> StackOverflowError on some configuration-related code.
>
> Here's a part of call-stack:
>
> Running on hadoop, using HADOOP_HOME=/opt/hadoop
> HADOOP_CONF_DIR=/opt/hadoop/conf
> MAHOUT-JOB: /home/sbos/gsoc/mahout/examples/target/mahout-examples-0.6-SNAPSHOT-job.jar
> 11/07/02 22:37:49 WARN driver.MahoutDriver: No hmmchunks.props found
> on classpath, will use command-line arguments only
> Exception in thread "main" java.lang.StackOverflowError
>       at java.io.UnixFileSystem.getBooleanAttributes0(Native Method)
>       at java.io.UnixFileSystem.getBooleanAttributes(UnixFileSystem.java:228)
>       at java.io.File.exists(File.java:733)
>       at sun.misc.URLClassPath$FileLoader.getResource(URLClassPath.java:999)
>       at sun.misc.URLClassPath$FileLoader.findResource(URLClassPath.java:966)
>       at sun.misc.URLClassPath.findResource(URLClassPath.java:146)
>       at java.net.URLClassLoader$2.run(URLClassLoader.java:385)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at java.net.URLClassLoader.findResource(URLClassLoader.java:382)
>       at java.lang.ClassLoader.getResource(ClassLoader.java:1003)
>       at java.lang.ClassLoader.getResource(ClassLoader.java:998)
>       at java.lang.ClassLoader.getResourceAsStream(ClassLoader.java:1193)
>       at javax.xml.parsers.SecuritySupport$4.run(SecuritySupport.java:96)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.xml.parsers.SecuritySupport.getResourceAsStream(SecuritySupport.java:89)
>       at javax.xml.parsers.FactoryFinder.findJarServiceProvider(FactoryFinder.java:250)
>       at javax.xml.parsers.FactoryFinder.find(FactoryFinder.java:223)
>       at javax.xml.parsers.DocumentBuilderFactory.newInstance(DocumentBuilderFactory.java:123)
>       at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1109)
>       at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:1093)
>       at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:1037)
>       at org.apache.hadoop.conf.Configuration.get(Configuration.java:415)
>       at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:168)
>       at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:159)
>       at org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:216)
>       at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:409)
>       at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:395)
>       at org.apache.hadoop.fs.FileSystem$Cache$Key.<init>(FileSystem.java:1418)
>       at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1319)
>       at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:226)
>       at org.apache.hadoop.fs.FsUrlConnection.connect(FsUrlConnection.java:45)
>       at org.apache.hadoop.fs.FsUrlConnection.getInputStream(FsUrlConnection.java:56)
>       at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown
> Source)
>       at org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown
> Source)
>
> and so on, and so on. Does it look like a bug in mahout/hadoop or my mistake?
>

Mime
View raw message