That makes sense.
I tried an alternate approach- i am using high level consumer and going
through Hadoop HDFS APIs and pushing data in HDFS.
I am not creating any jobs for that.
The only problem i am seeing here is that the consumer is designed to run
forever. Which means i need to find out how to close the HDFS file and kill
Is there any way to kill or close high level consumer gracefully?
I am running v0.7.0. I don't mind upgrading to higher version if that
allows me this kind of consumer handling.
On Thu, Jan 17, 2013 at 10:41 AM, Jun Rao <[EMAIL PROTECTED]> wrote: