ambari-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <>
Subject [jira] [Commented] (AMBARI-7119) log4j does not get used by hadoop as settings are present in
Date Mon, 20 Oct 2014 21:42:33 GMT


Hadoop QA commented on AMBARI-7119:

{color:red}-1 overall{color}.  Here are the results of testing the latest attachment
  against trunk revision .

    {color:green}+1 @author{color}.  The patch does not contain any @author tags.

    {color:red}-1 tests included{color}.  The patch doesn't appear to include any new or modified
                        Please justify why no new tests are needed for this patch.
                        Also please list what manual steps were performed to verify this patch.

    {color:green}+1 javac{color}.  The applied patch does not increase the total number of
javac compiler warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase the total number
of release audit warnings.

    {color:green}+1 core tests{color}.  The patch passed unit tests in ambari-server.

Test results:
Console output:

This message is automatically generated.

> log4j does not get used by hadoop as settings are present in
> -----------------------------------------------------------------------------
>                 Key: AMBARI-7119
>                 URL:
>             Project: Ambari
>          Issue Type: Bug
>            Reporter: Andrew Onischuk
>            Assignee: Dmitry Lysnichenko
>             Fix For: 1.7.0
>         Attachments: AMBARI-7119.patch.2, AMBARI-7119.patch.3
> PROBLEM: log4j settings made via Ambari update the log4j file but do not take
> any affect when restarting HDFS. It seems there are hardcoded settings in
> /usr/lib/hadoop/libexec/ such as this at line 221:
> HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.root.logger=$
> "
> BUSINESS IMPACT: Customers have to change core files or set environment
> variables explicitly by setting up a profile script
> STEPS TO REPRODUCE: Log in to Ambari and change log4j properties such that
> hadoop.root.logger=INFO,DRFA. The log4j file is updated in /etc/hadoop/conf.
> Restart the HDFS service. Do a ps -ef | grep <PID> for the namenode. The
> process shows duplicate entries for several properties and does not show the
> logging change. Here is the duplication and incorrect root logger setting seen
> locally in testing:
> hdfs 4304 1 14 07:26 ? 00:00:10 /usr/jdk64/jdk1.7.0_45/bin/java
> -Dproc_namenode -Xmx1024m
> -Dhadoop.log.dir=/var/log/hadoop/hdfs -Dhadoop.log.file=hadoop.log
> -Dhadoop.home.dir=/usr/lib/hadoop
> -Dhadoop.root.logger=INFO,console
> -Djava.library.path=:/usr/lib/hadoop/lib/native/Linux-
> amd64-64:/usr/lib/hadoop/lib/native -Dhadoop.policy.file=hadoop-policy.xml
> -Dhadoop.log.dir=/var/log/hadoop/hdfs
> -Dhadoop.home.dir=/usr/lib/hadoop
> -Dhadoop.root.logger=INFO,RFA -Djava.library.path=:/usr/lib/hadoop/lib/native
> /Linux-amd64-64:/usr/lib/hadoop/lib/native:/usr/lib/hadoop/lib/native/Linux-
> amd64-64:/usr/lib/hadoop/lib/native/Linux-amd64-64:/usr/lib/hadoop/lib/native
> -Dhadoop.policy.file=hadoop-policy.xml -server
> -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC
> -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:NewSize=100m
> -XX:MaxNewSize=50m -Xloggc:/var/log/hadoop/hdfs/gc.log-201407140726
> -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps
> -Xms1024m -Xmx1024m,DRFAS
> -Dhdfs.audit.logger=INFO,DRFAAUDIT -server -XX:ParallelGCThreads=8
> -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log
> -XX:NewSize=100m -XX:MaxNewSize=50m
> -Xloggc:/var/log/hadoop/hdfs/gc.log-201407140726 -verbose:gc
> -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms1024m
> -Xmx1024m,DRFAS
> -Dhdfs.audit.logger=INFO,DRFAAUDIT -server -XX:ParallelGCThreads=8
> -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log
> -XX:NewSize=100m -XX:MaxNewSize=50m
> -Xloggc:/var/log/hadoop/hdfs/gc.log-201407140726 -verbose:gc
> -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms1024m
> -Xmx1024m,DRFA
> ACTUAL BEHAVIOR: log4j changes made in Ambari do not persist in process. It
> seems there are values set in /usr/lib/hadoop/libexec/ that
> override no matter what. There is also duplication of settings, assuming this
> is from as well.
> EXPECTED BEHAVIOR: Settings made in Ambari should be persisted and used by the
> process.
> SUPPORT ANALYSIS: Support made changes to log4j in Ambari on a test cluster
> and they were not used

This message was sent by Atlassian JIRA

View raw message