You can use Kafka to store data on Hadoop via the Hadoop consumer in
contrib, and then use Talend or Pig to ETL it, before finally emitting the
ETL's records via the Hadoop producer in contrib.https://github.com/kafka-dev/kafka/tree/master/contribhttp://docs.hortonworks.com/CURRENT/index.htm#Data_Integration_Services_With_HDP/Using_Data_Integration_Services_Powered_By_Talend/Using_Talend.htm
Russell Jurney http://datasyndrome.com
On Jan 6, 2013, at 2:29 PM, David Arthur <[EMAIL PROTECTED]> wrote:
Storm has support for Kafka, if that's the sort of thing you're looking
for. Maybe you could describe your use case a bit more?
On Sunday, January 6, 2013, Guy Doulberg wrote:
I am looking for an ETL tool that can connect to kafka, as a consumer and
as a producer,
Have you heard of such a tool?