hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From comptech geeky <comptechge...@gmail.com>
Subject Re: Disc quota exceeded
Date Fri, 20 Jul 2012 16:34:11 GMT
After trying "df -kh". I got below result.

*bash-3.00$ df -kh*
*Filesystem             size   used  avail capacity  Mounted on*
*rpool/ROOT/sol10       916G    30G   668G     5%    /*
*/devices                 0K     0K     0K     0%    /devices*
*ctfs                     0K     0K     0K     0%    /system/contract*
*proc                     0K     0K     0K     0%    /proc*
*mnttab                   0K     0K     0K     0%    /etc/mnttab*
*swap                    31G   656K    31G     1%    /etc/svc/volatile*
*objfs                    0K     0K     0K     0%    /system/object*
*sharefs                  0K     0K     0K     0%    /etc/dfs/sharetab*
*/usr/lib/libc/libc_hwcap2.so.1*
*                       698G    30G   668G     5%    /lib/libc.so.1*
*fd                       0K     0K     0K     0%    /dev/fd*
*rpool/ROOT/sol10/var    20G    10G   9.7G    52%    /var*
*rpool/tmp               10G    10G     0K   100%    /tmp*
*swap                    31G    20K    31G     1%    /var/run*
*lvsaishdc3in0001data/data*
*                        32T    27T   2.4T    92%    /data*
*lvsaishdc3in0001data/data/b_apdpds*
*                       1.0T   8.5G  1016G     1%    /data/b_apdpds*
*lvsaishdc3in0001data/data/b_bids*
*                       100G    75G    25G    76%    /data/b_bids*
*lvsaishdc3in0001data/data/b_sbe*
*                       100G    51K   100G     1%    /data/b_sbe*
*lvsaishdc3in0001data/data/b_selling*
*                       500G   298G   202G    60%    /data/b_selling*
*lvsaishdc3in0001data/data/imk*
*                       3.0T   2.7T   293G    91%    /data/inbound/sq/imk*
*rpool/export           916G    23K   668G     1%    /export*
*rpool/export/home      175G   118G    57G    68%    /export/home*
*rpool                  916G    34K   668G     1%    /rpool*
*
*


On Fri, Jul 20, 2012 at 7:42 AM, kulkarni.swarnim@gmail.com <
kulkarni.swarnim@gmail.com> wrote:

> Seems to me like you might be just running out of disk space on one of the
> partitions. What does the output of "df -kh" say?
>
> Also, I just speculate that it might be your "/tmp" directory out of space
> because that is where hive tries to dump a bunch of log entries before it
> starts up. (/tmp/<user>/hive.log).
>
>
> On Fri, Jul 20, 2012 at 3:12 AM, comptech geeky <comptechgeeky@gmail.com>wrote:
>
>> Whenever I am typing Hive at the command prompt, I am getting the below
>> exception. What does that mean?
>> *
>> *
>> *$ bash*
>> *bash-3.00$ hive*
>> *Exception in thread "main" java.io.IOException: Disc quota exceeded*
>> *        at java.io.UnixFileSystem.createFileExclusively(Native Method)*
>> *        at java.io.File.checkAndCreate(File.java:1704)*
>> *        at java.io.File.createTempFile(File.java:1792)*
>> *        at org.apache.hadoop.util.RunJar.main(RunJar.java:115)*
>> *bash-3.00$*
>>
>> Any suggestions why is it happening?
>>
>
>
>
> --
> Swarnim
>

Mime
View raw message