Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
Kafka >> mail # user >> S3 Consumer


Copy link to this message
-
Re: S3 Consumer
I don't think anything exists like this in Kafka (or contrib), but it
would be a useful addition! Personally, I have written this exact thing
at previous jobs.

As for the Hadoop consumer, since there is a FileSystem implementation
for S3 in Hadoop, it should be possible. The Hadoop consumer works by
writing out data files containing the Kafka messages along side offset
files which contain the last offset read for each partition. If it is
re-consuming from zero each time you run it, it means it's not finding
the offset files from the previous run.

Having used it a bit, the Hadoop consumer is certainly an area that
could use improvement.

HTH,
David

On 12/27/12 4:41 AM, Pratyush Chandra wrote:
> Hi,
>
> I am looking for a S3 based consumer, which can write all the received
> events to S3 bucket (say every minute). Something similar to Flume HDFSSink
> http://flume.apache.org/FlumeUserGuide.html#hdfs-sink
> I have tried evaluating hadoop-consumer in contrib folder. But it seems to
> be more for offline processing, which will fetch everything from offset 0
> at once and replace it in S3 bucket.
> Any help would be appreciated ?
>
 
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB