hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Allen Wittenauer (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-6255) Create an rpm integration project
Date Mon, 14 Feb 2011 16:35:57 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-6255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12994356#comment-12994356
] 

Allen Wittenauer commented on HADOOP-6255:
------------------------------------------

> Allen, what location would you recommend for HADOOP_HOME in /usr path?

That's sort of the point of standardized locations.  HADOOP_HOME becomes superfluous.  In
fact, with a working ${BASH_SOURCE-0} and a few other changes in the shell commands, you can
run Hadoop 0.20 without setting $HADOOP_HOME/$HADOOP_CONF_DIR because hadoop-config.sh works
properly.

What should likely happen as part of this change, is hadoop-config.sh is modified to honor
any traditional HADOOP_HOME and HADOOP_CONF_DIR passed options and env settings.  If neither
exist, go look at the places defined in the packages.  Order should be something like:

1) --config location
2) HADOOP_CONF_DIR location
3) HADOOP_HOME/etc by new rules or if you want to be backward compat, HADOOP_HOME/conf
4) /etc/hadoop

Doing the above allows for a non-traditional location by setting HADOOP_HOME/HADOOP_CONF_DIR
just as one would do so today. 

The big question is whether or not you want to honor ${BASH_SOURCE}/../etc or ${BASH_SOURCE}/../conf
similar to how it works in the current releases.  My thought is no, as we provide more than
enough hooks to override with just the four above.

One of the tricky issues is what to do about hadoop-config.sh itself.  The "new standard"
is typically to compile in the logic into commands and then provide an "app-config" shell
script in /usr/bin that provides the well known locations to external apps (which is really
the only "new" bit).  This is probably the approach we should take.  At build time, hadoop-config.sh
or equiv gets sucked into the various shell commands including a new hadoop-config executable
so that HBase and others get the info they need to execute properly. Another choice is to
include hooks into pkgconfig, but I don't think that's as fluid as we really want here.

> Create an rpm integration project
> ---------------------------------
>
>                 Key: HADOOP-6255
>                 URL: https://issues.apache.org/jira/browse/HADOOP-6255
>             Project: Hadoop Common
>          Issue Type: New Feature
>    Affects Versions: 0.20.100
>            Reporter: Owen O'Malley
>            Assignee: Eric Yang
>             Fix For: 0.20.100
>
>         Attachments: HADOOP-6255-branch-0.20-security.patch, HADOOP-6255.patch, deployment.pdf
>
>
> We should be able to create RPMs for Hadoop releases.

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Mime
View raw message