hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Arun C Murthy (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-256) Implement a C api for hadoop dfs
Date Mon, 29 May 2006 03:23:30 GMT
    [ http://issues.apache.org/jira/browse/HADOOP-256?page=comments#action_12413660 ] 

Arun C Murthy commented on HADOOP-256:


  Appreciate your feedback... some minor queries/responses:

    a) I'm working on getting the whole thing integrated with build.xml/ant and also about
putting all outputs in the top-level build directory. librecordio led me down the wrong garden
    b) Can you please elaborate on the issues you had with LDFLAGS? On my system (Gentoo GNU/Linux
- make/makeclean works for me i.e. there are no issues building it.  I'm relying
on $JAVA_HOME in my Makefile (it's set to /opt/sun-jdk- on my box - afaik it's the
recommended way?) to point me libjvm.so. This isn't prefect (e.g. $(JAVA_HOME)/jre/lib/i386/server
won't work on non i386 architectures??) Is there a better alternative?
       Since I wasn't clear which system path to dump the .so I'm building (any suggestions?),
I just used LD_PRELOAD to run my executable (hdfs_test) - admit I should have clearly pointed
this out. 
       For the same reason I had to resort to --rpath tricks to get TestDFSCIO to work (i.e.
both hdfs_read & hdfs_write executables have an --rpath for libhdfs.so.1). This was very
necessary for getting the hdfs_{read|write} to work with map/reduce. Appreciate any suggestions...
    c) Does it make sense to just copy the executables from $(HADOOP_HOME)/libhdfs to /tmp/TestDFSCIO
from the client (initialization phase) on which the tests are initiated? Admit this completely
slipped my mind, glaringly. My bad. 


> Implement a C api for hadoop dfs
> --------------------------------
>          Key: HADOOP-256
>          URL: http://issues.apache.org/jira/browse/HADOOP-256
>      Project: Hadoop
>         Type: New Feature

>   Components: dfs
>     Reporter: Arun C Murthy
>  Attachments: libhdfs.patch
> Implement a C api for hadoop dfs to ease talking to hadoop's dfs from native C/C++ applications.

This message is automatically generated by JIRA.
If you think it was sent incorrectly contact one of the administrators:
For more information on JIRA, see:

View raw message