hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Srikanth Sundarrajan (JIRA)" <j...@apache.org>
Subject [jira] Updated: (HDFS-481) Bug Fixes + HdfsProxy to use proxy user to impresonate the real user
Date Tue, 20 Apr 2010 12:07:53 GMT

     [ https://issues.apache.org/jira/browse/HDFS-481?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Srikanth Sundarrajan updated HDFS-481:
--------------------------------------

    Attachment: HDFS-481-bp-y20s.patch

Incremental back port to fix broken unit tests in y20.1xx & y20.101. Tests are broken
due to 

* Missing super user setup when Mini DFS Cluster starts
* Missing src/test/resources folder
* UserGroupInformation class depending on krb5.conf in the system. (Bypassing that through
krb5.conf in ${hadoop.core}/src/test  - contrib/hdfsproxy/build.xml change)  

This patch needs to be applied incrementally over the HDFS-481-NEW.patch.

> Bug Fixes + HdfsProxy to use proxy user to impresonate the real user
> --------------------------------------------------------------------
>
>                 Key: HDFS-481
>                 URL: https://issues.apache.org/jira/browse/HDFS-481
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: contrib/hdfsproxy
>    Affects Versions: 0.21.0
>            Reporter: zhiyong zhang
>            Assignee: Srikanth Sundarrajan
>             Fix For: 0.22.0
>
>         Attachments: HDFS-481-bp-y20.patch, HDFS-481-bp-y20.patch, HDFS-481-bp-y20s.patch,
HDFS-481-bp-y20s.patch, HDFS-481-bp-y20s.patch, HDFS-481-NEW.patch, HDFS-481.out, HDFS-481.patch,
HDFS-481.patch, HDFS-481.patch, HDFS-481.patch, HDFS-481.patch, HDFS-481.patch, HDFS-481.patch,
HDFS-481.patch
>
>
> Bugs:
> 1. hadoop-version is not recognized if run ant command from src/contrib/ or from src/contrib/hdfsproxy
 
> If running ant command from $HADOOP_HDFS_HOME, hadoop-version will be passed to contrib's
build through subant. But if running from src/contrib or src/contrib/hdfsproxy, the hadoop-version
will not be recognized. 
> 2. LdapIpDirFilter.java is not thread safe. userName, Group & Paths are per request
and can't be class members.
> 3. Addressed the following StackOverflowError. 
> ERROR [org.apache.catalina.core.ContainerBase.[Catalina].[localh
> ost].[/].[proxyForward]] Servlet.service() for servlet proxyForward threw exception
> java.lang.StackOverflowError
>         at org.apache.catalina.core.ApplicationHttpRequest.getAttribute(ApplicationHttpR
> equest.java:229)
>      This is due to when the target war (/target.war) does not exist, the forwarding
war will forward to its parent context path /, which defines the forwarding war itself. This
cause infinite loop.  Added "HDFS Proxy Forward".equals(dstContext.getServletContextName()
in the if logic to break the loop.
> 4. Kerberos credentials of remote user aren't available. HdfsProxy needs to act on behalf
of the real user to service the requests

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message