hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Martin Traverso (JIRA)" <j...@apache.org>
Subject [jira] Updated: (HADOOP-2845) dfsadmin disk utilization report on Solaris is wrong
Date Thu, 21 Feb 2008 05:50:45 GMT

     [ https://issues.apache.org/jira/browse/HADOOP-2845?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel

Martin Traverso updated HADOOP-2845:

    Attachment: HADOOP-2845-1.patch

> dfsadmin disk utilization report on Solaris is wrong
> ----------------------------------------------------
>                 Key: HADOOP-2845
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2845
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: fs
>    Affects Versions: 0.16.0
>            Reporter: Martin Traverso
>             Fix For: 0.17.0
>         Attachments: HADOOP-2845-1.patch, HADOOP-2845.patch
> dfsadmin reports 2x disk utilization on some platforms (Solaris, MacOS). The reason for
this is that org.apache.hadoop.fs.DU is relying on du's default block size when reporting
sizes and assuming they are 1024 byte blocks. This works fine on Linux, but du Solaris and
MacOS uses 512-byte blocks to report disk usage.
> DU should use "du -sk" instead of "du -s" to force the command to report sizes based
on 1024 byte blocks.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message