臧冬松 2011-12-23, 09:47
I'd recommend trying the HTTP api (webhdfs - http://hortonworks.com/webhdfs-–-http-rest-access-to-hdfs/) from your C++ application or FUSE-DFS (the C api - http://hadoop.apache.org/common/docs/r0.20.205.0/libhdfs.html).
PS: Not sure why webhdfs docs on hadoop.apache.org aren't showing up. Maybe they weren't in 0.20.205 and are only fully formed in hadoop-1.0 which just got released. I'll wait for Matt to update the site.
On Dec 23, 2011, at 1:47 AM, 臧冬松 wrote:
> In my application, I have some big c++ codes. And now I want to store files in HDFS, and use the c++ programs to read/write from/to HDFS.
> I wonder what's the best way(efficient,high throughput) to do this.
> I know there's a libhdfs which is based on JNI, and a conrib code of Thrift (which is removed from trunck now).
> But if I really care about performance in reading and writing data, is there other ways to do this?
> I notice that there's a JIRA(https://issues.apache.org/jira/browse/HDFS-2478) using protobuf as the IPC protocal, is it a good idea that we can write a client lib in C++ using the new protobuf IPC?