Return-Path: X-Original-To: apmail-mahout-user-archive@www.apache.org Delivered-To: apmail-mahout-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 24486DAE0 for ; Sat, 1 Sep 2012 15:40:05 +0000 (UTC) Received: (qmail 63049 invoked by uid 500); 1 Sep 2012 15:40:02 -0000 Delivered-To: apmail-mahout-user-archive@mahout.apache.org Received: (qmail 63015 invoked by uid 500); 1 Sep 2012 15:40:02 -0000 Mailing-List: contact user-help@mahout.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@mahout.apache.org Delivered-To: mailing list user@mahout.apache.org Received: (qmail 63007 invoked by uid 99); 1 Sep 2012 15:40:02 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 01 Sep 2012 15:40:02 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FSL_RCVD_USER,HTML_MESSAGE,NORMAL_HTTP_TO_IP,RCVD_IN_DNSWL_LOW,SPF_PASS,WEIRD_PORT X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of ted.dunning@gmail.com designates 74.125.82.50 as permitted sender) Received: from [74.125.82.50] (HELO mail-wg0-f50.google.com) (74.125.82.50) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 01 Sep 2012 15:39:56 +0000 Received: by wgbds11 with SMTP id ds11so2305616wgb.7 for ; Sat, 01 Sep 2012 08:39:34 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :content-type; bh=DJMQeCwMyHADDJl5HS+TUtDze9WoMdk6DReL5ImIE5g=; b=ucA2ZEm3KqjpV0wCSOA2XyJpZkytA1ySIZfXk72yL4oP/yq471im4NHMdozPCBJZRw 60DwV8mNLrFMyk/tSNAi02wrtZpZe022wThv6Awnh1JHuipNozFsoDjp8Pwg9ZLnpSyK KxEDhi1tDITPvzQX0g0655Al4mDyItJZFPkxfszv/5+boTHCBouD8fOHyGRLlEAvBT6N 01bhK/jjZXuYBh9oLfrbl6XQz96KK2TwMmzbdsU0ttOh8pS4g6h9LH7bZMtz0wMlZwe6 OxMHZ1mOkUE77B+eml1MvY6Zfb218NnPhaZ5/ODoTw3LkDYPJwnqQwQOf4dpTqOOIoSL wO6A== Received: by 10.180.81.165 with SMTP id b5mr11766311wiy.17.1346513974823; Sat, 01 Sep 2012 08:39:34 -0700 (PDT) MIME-Version: 1.0 Received: by 10.223.63.211 with HTTP; Sat, 1 Sep 2012 08:39:04 -0700 (PDT) In-Reply-To: <9FA1ECCA-C382-4FB4-B098-37438DCB623A@gmail.com> References: <9FA1ECCA-C382-4FB4-B098-37438DCB623A@gmail.com> From: Ted Dunning Date: Sat, 1 Sep 2012 11:39:04 -0400 Message-ID: Subject: Re: SSVD error To: user@mahout.apache.org Content-Type: multipart/alternative; boundary=f46d044304ee278a5704c8a5b7f0 X-Virus-Checked: Checked by ClamAV on apache.org --f46d044304ee278a5704c8a5b7f0 Content-Type: text/plain; charset=UTF-8 Regardless of confusion between k and p (I was confused as well) you still can't set the sum to more than the minimum size of your data. Here you have set it larger. And it breaks. On Sat, Sep 1, 2012 at 11:09 AM, Pat Ferrel wrote: > Oh, sorry, below I meant to say k (the number to reduce to) not p. > > In both cases p = 1, the first case k = 20, the second case k =100 > > Also the first error does seem to be running with local hadoop. The error > is from looking for a temp file that does exist in the file system, but not > in the hadoop tmp based files. > > > On Sep 1, 2012, at 7:53 AM, Ted Dunning wrote: > > With 57 crawled docs, you can't reasonably set p > 57. That is your second > error. > > On Sat, Sep 1, 2012 at 10:32 AM, Pat Ferrel wrote: > > > I have a small data set that I am using in local mode for debugging > > purposes. The data is 57 crawled docs with something like 2200 terms. I > run > > this through seq2sparse, then my own cloned version of rowid to get a > > distributed row matrix, then into SSVD. I realize this is not a > production > > environment, but you need to debug somewhere and single threaded > execution > > is ideal for debugging. As I said this works in hadoop clustered mode. > > > > The error looks like some code is expecting hdfs to be running, no? Here > > is the exception stack from the ide with p = 20: > > > > 12/09/01 07:22:55 WARN mapred.LocalJobRunner: job_local_0002 > > java.io.FileNotFoundException: File > > > /tmp/hadoop-pat/mapred/local/archive/6590995089539988730_1587570556_37122331/file/Users/pat/Projects/big-data/b/ssvd/Q-job/R-m-00000 > > does not exist. > > at > > > org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:371) > > at > > > org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:245) > > at > > > org.apache.mahout.common.iterator.sequencefile.SequenceFileDirValueIterator.(SequenceFileDirValueIterator.java:92) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.BtJob$BtMapper.setup(BtJob.java:219) > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370) > > at > > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:212) > > Exception in thread "main" java.io.IOException: Bt job unsuccessful. > > at > > org.apache.mahout.math.hadoop.stochasticsvd.BtJob.run(BtJob.java:609) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.SSVDSolver.run(SSVDSolver.java:397) > > at > > > com.finderbots.analysis.AnalysisPipeline.SSVDTransformAndBack(AnalysisPipeline.java:257) > > at com.finderbots.analysis.AnalysisJob.run(AnalysisJob.java:20) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > Disconnected from the target VM, address: '127.0.0.1:54588', transport: > > 'socket' > > at com.finderbots.analysis.AnalysisJob.main(AnalysisJob.java:34) > > > > Process finished with exit code 1 > > > > With p=100-200 I get the following: > > > > 12/09/01 07:30:33 ERROR common.IOUtils: new m can't be less than n > > java.lang.IllegalArgumentException: new m can't be less than n > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.GivensThinSolver.adjust(GivensThinSolver.java:109) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.QRFirstStep.cleanup(QRFirstStep.java:233) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.QRFirstStep.close(QRFirstStep.java:89) > > at org.apache.mahout.common.IOUtils.close(IOUtils.java:128) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.QJob$QMapper.cleanup(QJob.java:158) > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370) > > at > > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:212) > > 12/09/01 07:30:33 WARN mapred.LocalJobRunner: job_local_0001 > > java.lang.IllegalArgumentException: new m can't be less than n > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.GivensThinSolver.adjust(GivensThinSolver.java:109) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.QRFirstStep.cleanup(QRFirstStep.java:233) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.qr.QRFirstStep.close(QRFirstStep.java:89) > > at org.apache.mahout.common.IOUtils.close(IOUtils.java:128) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.QJob$QMapper.cleanup(QJob.java:158) > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370) > > at > > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:212) > > Exception in thread "main" java.io.IOException: Q job unsuccessful. > > at > > org.apache.mahout.math.hadoop.stochasticsvd.QJob.run(QJob.java:230) > > at > > > org.apache.mahout.math.hadoop.stochasticsvd.SSVDSolver.run(SSVDSolver.java:376) > > at > > > com.finderbots.analysis.AnalysisPipeline.SSVDTransformAndBack(AnalysisPipeline.java:257) > > at com.finderbots.analysis.AnalysisJob.run(AnalysisJob.java:20) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > at com.finderbots.analysis.AnalysisJob.main(AnalysisJob.java:34) > > Disconnected from the target VM, address: '127.0.0.1:54614', transport: > > 'socket' > > > > Process finished with exit code 1 > > > > > > > > > > On Aug 31, 2012, at 4:21 PM, Dmitriy Lyubimov wrote: > > > > Perhaps if you give more info about the stack etc. i might get a > > better idea though > > > > On Fri, Aug 31, 2012 at 4:19 PM, Dmitriy Lyubimov > > wrote: > >> I am not sure, i haven't used it that way. > >> > >> I know it works fully distributed AND when embedded with local job > >> tracker (e.g. its tests are basically MR jobs with "local" job > >> tracker) which probably is not the same as Mahout local mode. "local" > >> job tracker is not good for much though: thus it doesn't use even > >> multicore parallelism as it doesn't support multiple reducers, so this > >> code is kind of for a real cluster really, pragmatically. There's also > >> Ted's implementation of non-distributed SSVD in Mahout which does not > >> require Hadoop dependencies but it is a different api with no PCA > >> option (not sure about power iterations). > >> > >> I am not sure why this very particular error appears in your setup. > >> > >> On Fri, Aug 31, 2012 at 3:02 PM, Pat Ferrel > > wrote: > >>> Running on the local file system inside IDEA with MAHOUT_LOCAL set and > > performing an SSVD I get the error below. Notice that R-m-00000 exists in > > the local file system and running it outside the debugger in > pseudo-cluster > > mode with HDFS works. Does SSVD work in local mode? > >>> > >>> java.io.FileNotFoundException: File > > > /tmp/hadoop-pat/mapred/local/archive/5543644668644532045_1587570556_2120541978/file/Users/pat/Projects/big-data/b/ssvd/Q-job/R-m-00000 > > does not exist. > >>> > >>> Maclaurin:big-data pat$ ls -al b/ssvd/Q-job/ > >>> total 72 > >>> drwxr-xr-x 10 pat staff 340 Aug 31 13:35 . > >>> drwxr-xr-x 4 pat staff 136 Aug 31 13:35 .. > >>> -rw-r--r-- 1 pat staff 80 Aug 31 13:35 .QHat-m-00000.crc > >>> -rw-r--r-- 1 pat staff 28 Aug 31 13:35 .R-m-00000.crc > >>> -rw-r--r-- 1 pat staff 8 Aug 31 13:35 ._SUCCESS.crc > >>> -rw-r--r-- 1 pat staff 12 Aug 31 13:35 .part-m-00000.deflate.crc > >>> -rwxrwxrwx 1 pat staff 9154 Aug 31 13:35 QHat-m-00000 > >>> -rwxrwxrwx 1 pat staff 2061 Aug 31 13:35 R-m-00000 > >>> -rwxrwxrwx 1 pat staff 0 Aug 31 13:35 _SUCCESS > >>> -rwxrwxrwx 1 pat staff 8 Aug 31 13:35 part-m-00000.deflate > >>> > > > > > > --f46d044304ee278a5704c8a5b7f0--