Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Plain View
Kafka, mail # user - Payload size exception


+
Bo Sun 2013-01-29, 08:18
Copy link to this message
-
Re: Payload size exception
Jay Kreps 2013-01-29, 16:05
There is a setting that controls the maximum message size. This is to
ensure the messages can be read on the server and by all consumers without
running out of memory or exceeding the consumer fetch size. In 0.7.x this
setting is controlled by the broker configuration max.message.size.

-Jay
On Tue, Jan 29, 2013 at 12:18 AM, Bo Sun <[EMAIL PROTECTED]> wrote:

> hi all
> i've got a exception .
> kafka.common.MessageSizeTooLargeException: payload size of 1772597 larger
> than 1000000
>         at
>
> kafka.message.ByteBufferMessageSet.verifyMessageSize(ByteBufferMessageSet.scala:93)
>         at kafka.producer.SyncProducer.send(SyncProducer.scala:122)
>         at
>
> kafka.producer.ProducerPool$$anonfun$send$1.apply$mcVI$sp(ProducerPool.scala:114)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57)
>         at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43)
>         at kafka.producer.ProducerPool.send(ProducerPool.scala:100)
>         at kafka.producer.Producer.zkSend(Producer.scala:140)
>         at kafka.producer.Producer.send(Producer.scala:99)
>         at kafka.javaapi.producer.Producer.send(Producer.scala:103)
>
> i dont know why. plz help me .thanks
>

 
+
Neha Narkhede 2013-01-29, 16:35
+
S Ahmed 2013-01-29, 16:57
+
Xavier Stevens 2013-01-29, 17:21
+
Neha Narkhede 2013-01-29, 17:28
+
S Ahmed 2013-01-29, 19:00
+
Jay Kreps 2013-01-29, 18:44