Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Kafka, mail # user - Payload size exception


Copy link to this message
-
Re: Payload size exception
Jay Kreps 2013-01-29, 16:05
There is a setting that controls the maximum message size. This is to
ensure the messages can be read on the server and by all consumers without
running out of memory or exceeding the consumer fetch size. In 0.7.x this
setting is controlled by the broker configuration max.message.size.

-Jay
On Tue, Jan 29, 2013 at 12:18 AM, Bo Sun <[EMAIL PROTECTED]> wrote:

> hi all
> i've got a exception .
> kafka.common.MessageSizeTooLargeException: payload size of 1772597 larger
> than 1000000
>         at
>
> kafka.message.ByteBufferMessageSet.verifyMessageSize(ByteBufferMessageSet.scala:93)
>         at kafka.producer.SyncProducer.send(SyncProducer.scala:122)
>         at
>
> kafka.producer.ProducerPool$$anonfun$send$1.apply$mcVI$sp(ProducerPool.scala:114)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57)
>         at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43)
>         at kafka.producer.ProducerPool.send(ProducerPool.scala:100)
>         at kafka.producer.Producer.zkSend(Producer.scala:140)
>         at kafka.producer.Producer.send(Producer.scala:99)
>         at kafka.javaapi.producer.Producer.send(Producer.scala:103)
>
> i dont know why. plz help me .thanks
>