Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
Flume >> mail # user >> unsubscribe


On Fri, Sep 27, 2013 at 1:43 PM, Paul Chavez <
[EMAIL PROTECTED]> wrote:

> Thanks for the update. I remember I had a similar situation now, except
> that I had the transactionCapacity lower than the batch size for the sink.
> I guess having them exactly the same is not optimal either.****
>
> ** **
>
> -Paul****
>
> ** **
>
> ** **
>
> *From:* Cameron Wellock [mailto:[EMAIL PROTECTED]]
> *Sent:* Friday, September 27, 2013 7:55 AM
> *To:* [EMAIL PROTECTED]
> *Subject:* Re: Unable to put batch on required channel****
>
> ** **
>
> Final update, in case anyone ever has a similar problem: increasing the
> transactionCapacity to a low multiple of the batch size (say 5x batch size)
> seems to have fixed the problem, at least for the moment. ****
>
> ** **
>
> Cameron****
>
> ** **
>
> On Thu, Sep 26, 2013 at 12:22 PM, Cameron Wellock <
> [EMAIL PROTECTED]> wrote:****
>
> Hi Paul, thanks for your thoughts. The sink does not complain--at all--and
> there are no relevant errors in the logs on the datanodes. I haven't waited
> to see if flume recovers after the other write stops, as I took the error
> messages at face value and restarted flume. I will try that today, time
> permitting, and I'll let you know what happens.****
>
> ** **
>
> Thanks again,****
>
> Cameron****
>
> ** **
>
> On Thu, Sep 26, 2013 at 12:07 PM, Paul Chavez <
> [EMAIL PROTECTED]> wrote:****
>
> Is the HDFS sink reporting any issues writing to the cluster? If you leave
> it alone or wait until the other application stops writing will flume
> recover?****
>
>  ****
>
> SpoolDir is a good source if the write performance to HDFS is variable as
> the files in the spool directory will just sit and wait until the flume
> channel has space again. Another option may be to add another HDFS sink or
> two pulling from the same channel, but from what you are saying this may
> not increase performance.****
>
>  ****
>
> Hope that helps,****
>
> Paul Chavez****
>
>  ****
>
>  ****
>
> *From:* Cameron Wellock [mailto:[EMAIL PROTECTED]]
> *Sent:* Thursday, September 26, 2013 8:37 AM
> *To:* [EMAIL PROTECTED]
> *Subject:* Unable to put batch on required channel****
>
>  ****
>
> Hello world,****
>
>  ****
>
> I've been trying to set up a test instance of flume and have been stymied
> by recurring failures. I'm trying to use a single flume agent moving about
> 200G of data from a spooldir into a very small hadoop cluster (3 nodes). If
> flume is the only thing writing to HDFS, everything works fine, but as soon
> as another application starts writing data into the cluster HDFS slows down
> and flume barfs with an "unable to put batch on required channel" exception.
> ****
>
>  ****
>
> I have tried all kinds of configuration changes, to no avail. I have tried
> memory channels, file channels, small batch sizes (down to 50), large batch
> sizes (up to 20000), increasing timeouts, increasing channel capacity (up
> to 150 million), you name it. Sooner or later (usually 5-10 minutes after
> restart) flume comes to a halt. This is especially vexing considering that
> it's copying from a file to a file--there are no realtime requirements that
> might reasonably lead to a full channel in other circumstances. Anybody
> have any advice? Insights? Wild guesses? Outright lies?****
>
>  ****
>
> Below are two exceptions from the log, one from a memory channel
> configuration, one from a file channel configuration, and below that is the
> most recent configuration file used. Absolutely any suggestions would be
> appreciated.****
>
>  ****
>
> Thanks,****
>
> Cameron****
>
>  ****
>
>  ****
>
> 25 Sep 2013 21:05:12,262 ERROR [pool-5-thread-1]
> (org.apache.flume.source.SpoolDirectorySource$Spool****
>
> DirectoryRunnable.run:195)  - FATAL: Spool Directory source r1: {
> spoolDir: /var/nrelate/flume-spool****
>
>  }: Uncaught exception in SpoolDirectorySource thread. Restart or
> reconfigure Flume to continue proc****
>
> essing.****
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB