Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 61B561097A for ; Fri, 24 Jan 2014 14:48:27 +0000 (UTC) Received: (qmail 11981 invoked by uid 500); 24 Jan 2014 14:48:20 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 11809 invoked by uid 500); 24 Jan 2014 14:48:19 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 11802 invoked by uid 99); 24 Jan 2014 14:48:19 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 24 Jan 2014 14:48:19 +0000 X-ASF-Spam-Status: No, hits=2.2 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (nike.apache.org: 216.145.54.172 is neither permitted nor denied by domain of tep@yahoo-inc.com) Received: from [216.145.54.172] (HELO mrout2.yahoo.com) (216.145.54.172) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 24 Jan 2014 14:48:13 +0000 Received: from BF1-EX10-CAHT03.y.corp.yahoo.com (bf1-ex10-caht03.corp.bf1.yahoo.com [10.74.209.58]) by mrout2.yahoo.com (8.14.4/8.14.4/y.out) with ESMTP id s0OElMn6018739 (version=TLSv1/SSLv3 cipher=AES128-SHA bits=128 verify=FAIL) for ; Fri, 24 Jan 2014 06:47:23 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=yahoo-inc.com; s=cobra; t=1390574843; bh=gPbBUcSvAIkG6B62+y5tb8a3zGkA37jmT4ceF97+SaI=; h=Message-ID:Date:From:MIME-Version:To:Subject:Content-Type; b=roJGsoPPeYag/6zbJSaz1WTUgdqrd7ZfkNe7I6bsE/PAqECw4CakGaeNcx3Tsyesx aDafMH6MHlq8QFCIcCMNmjuzpgTtnd6ZUXeQk3Rr3KVFCoyODGZjawGBw0SrZo0j4Z i7uRNj92O18mZrOQvaiPVOSa0qTQuQd6yUO9wgKc= Received: from [10.78.36.114] (10.78.36.114) by owa.corp.yahoo.com (10.74.209.170) with Microsoft SMTP Server (TLS) id 14.3.174.1; Fri, 24 Jan 2014 09:47:20 -0500 Message-ID: <52E27CF9.6030109@yahoo-inc.com> Date: Fri, 24 Jan 2014 15:47:21 +0100 From: Tim Potter User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.9; rv:24.0) Gecko/20100101 Thunderbird/24.2.0 MIME-Version: 1.0 To: Subject: No space left on device during merge. Content-Type: multipart/alternative; boundary="------------080508090501000106060104" X-Originating-IP: [10.78.36.114] X-Milter-Version: master.31+4-gbc07cd5+ X-CLX-ID: 574843002 X-Virus-Checked: Checked by ClamAV on apache.org --------------080508090501000106060104 Content-Type: text/plain; charset="ISO-8859-1"; format=flowed Content-Transfer-Encoding: 7bit Hi, I'm getting the below error while trying to sort a lot of data with Hadoop. I strongly suspect the node the merge is on is running out of local disk space. Assuming this is the case, is there any way to get around this limitation considering I can't increase the local disk space available on the nodes? Like specify sort/merge parameters or similar. Thanks, Tim. 2014-01-24 10:02:36,267 INFO [main] org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor [.lzo_deflate] 2014-01-24 10:02:36,280 INFO [main] org.apache.hadoop.mapred.Merger: Down to the last merge-pass, with 100 segments left of total size: 642610678884 bytes 2014-01-24 10:02:36,281 ERROR [main] org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:XXXXXX (auth:XXXXXX) cause:org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error in shuffle in OnDiskMerger - Thread to merge on-disk map-outputs 2014-01-24 10:02:36,282 WARN [main] org.apache.hadoop.mapred.YarnChild: Exception running child : org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error in shuffle in OnDiskMerger - Thread to merge on-disk map-outputs at org.apache.hadoop.mapreduce.task.reduce.Shuffle.run(Shuffle.java:167) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:371) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:158) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1284) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:153) Caused by: org.apache.hadoop.fs.FSError: java.io.IOException: No space left on device at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:213) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.write(BufferedOutputStream.java:126) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:54) at java.io.DataOutputStream.write(DataOutputStream.java:107) at org.apache.hadoop.mapred.IFileOutputStream.write(IFileOutputStream.java:88) at org.apache.hadoop.io.compress.BlockCompressorStream.compress(BlockCompressorStream.java:150) at org.apache.hadoop.io.compress.BlockCompressorStream.finish(BlockCompressorStream.java:140) at org.apache.hadoop.io.compress.BlockCompressorStream.write(BlockCompressorStream.java:99) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:54) at java.io.DataOutputStream.write(DataOutputStream.java:107) at org.apache.hadoop.mapred.IFile$Writer.append(IFile.java:249) at org.apache.hadoop.mapred.Merger.writeFile(Merger.java:200) at org.apache.hadoop.mapreduce.task.reduce.MergeManager$OnDiskMerger.merge(MergeManager.java:572) at org.apache.hadoop.mapreduce.task.reduce.MergeThread.run(MergeThread.java:94) Caused by: java.io.IOException: No space left on device at java.io.FileOutputStream.writeBytes(Native Method) at java.io.FileOutputStream.write(FileOutputStream.java:318) at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:211) ... 14 more 2014-01-24 10:02:36,284 INFO [main] org.apache.hadoop.mapred.Task: Runnning cleanup for the task --------------080508090501000106060104 Content-Type: text/html; charset="ISO-8859-1" Content-Transfer-Encoding: 7bit
Hi,
  I'm getting the below error while trying to sort a lot of data with Hadoop.

I strongly suspect the node the merge is on is running out of local disk space. Assuming this is the case, is there any way
to get around this limitation considering I can't increase the local disk space available on the nodes?  Like specify sort/merge parameters or similar.

Thanks,
  Tim.

2014-01-24 10:02:36,267 INFO [main] org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor [.lzo_deflate]
2014-01-24 10:02:36,280 INFO [main] org.apache.hadoop.mapred.Merger: Down to the last merge-pass, with 100 segments left of total size: 642610678884 bytes
2014-01-24 10:02:36,281 ERROR [main] org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:XXXXXX (auth:XXXXXX) cause:org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error in shuffle in OnDiskMerger - Thread to merge on-disk map-outputs
2014-01-24 10:02:36,282 WARN [main] org.apache.hadoop.mapred.YarnChild: Exception running child : org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error in shuffle in OnDiskMerger - Thread to merge on-disk map-outputs
	at org.apache.hadoop.mapreduce.task.reduce.Shuffle.run(Shuffle.java:167)
	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:371)
	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:158)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1284)
	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:153)
Caused by: org.apache.hadoop.fs.FSError: java.io.IOException: No space left on device
	at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:213)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.write(BufferedOutputStream.java:126)
	at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:54)
	at java.io.DataOutputStream.write(DataOutputStream.java:107)
	at org.apache.hadoop.mapred.IFileOutputStream.write(IFileOutputStream.java:88)
	at org.apache.hadoop.io.compress.BlockCompressorStream.compress(BlockCompressorStream.java:150)
	at org.apache.hadoop.io.compress.BlockCompressorStream.finish(BlockCompressorStream.java:140)
	at org.apache.hadoop.io.compress.BlockCompressorStream.write(BlockCompressorStream.java:99)
	at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:54)
	at java.io.DataOutputStream.write(DataOutputStream.java:107)
	at org.apache.hadoop.mapred.IFile$Writer.append(IFile.java:249)
	at org.apache.hadoop.mapred.Merger.writeFile(Merger.java:200)
	at org.apache.hadoop.mapreduce.task.reduce.MergeManager$OnDiskMerger.merge(MergeManager.java:572)
	at org.apache.hadoop.mapreduce.task.reduce.MergeThread.run(MergeThread.java:94)
Caused by: java.io.IOException: No space left on device
	at java.io.FileOutputStream.writeBytes(Native Method)
	at java.io.FileOutputStream.write(FileOutputStream.java:318)
	at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:211)
	... 14 more

2014-01-24 10:02:36,284 INFO [main] org.apache.hadoop.mapred.Task: Runnning cleanup for the task
--------------080508090501000106060104--