Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Plain View
Kafka, mail # user - java.io.IOException: Broken pipe


+
Yonghui Zhao 2013-03-15, 09:33
+
Neha Narkhede 2013-03-15, 13:06
+
Yonghui Zhao 2013-03-18, 14:26
+
Neha Narkhede 2013-03-19, 06:06
+
Yonghui Zhao 2013-03-19, 13:16
+
Neha Narkhede 2013-03-19, 13:43
+
Yonghui Zhao 2013-03-20, 08:36
Copy link to this message
-
Re: java.io.IOException: Broken pipe
Neha Narkhede 2013-03-20, 14:58
From the consumer logs, it seems senseidb is interrupting the simple
consumer thread. This causes the socket to close which then shows up as a
broken pipe on the server. I don't know senseidb to say if this thread
interruption makes sense. But there are better ways to close the consumer
properly. I would post this question on sensei db mailing list.

Thanks,
Neha
On Wed, Mar 20, 2013 at 1:35 AM, Yonghui Zhao <[EMAIL PROTECTED]> wrote:

> Thanks Neha,
>
> After enable INFO log in consumer,
> I find 2 exceptions in consumer side, any idea?
>
>
>
> 2013/03/20 14:52:00.585 INFO [SimpleConsumer] [] Reconnect in multifetch
> due to socket error:
> java.nio.channels.ClosedChannelException
>     at
> sun.nio.ch.SocketChannelImpl.ensureReadOpen(SocketChannelImpl.java:120)
>     at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:163)
>     at kafka.utils.Utils$.read(Utils.scala:538)
>     at
>
> kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:67)
>     at kafka.network.Receive$class.readCompletely(Transmission.scala:55)
>     at
>
> kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
>     at kafka.consumer.SimpleConsumer.getResponse(SimpleConsumer.scala:177)
>     at
> kafka.consumer.SimpleConsumer.liftedTree2$1(SimpleConsumer.scala:117)
>     at kafka.consumer.SimpleConsumer.multifetch(SimpleConsumer.scala:115)
>     at kafka.consumer.FetcherRunnable.run(FetcherRunnable.scala:60)
>
>
> 2013/03/20 14:52:03.678 INFO [SimpleConsumer] [] Reconnect in multifetch
> due to socket error:
> java.nio.channels.ClosedByInterruptException
>     at
>
> java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:184)
>     at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:270)
>     at kafka.utils.Utils$.read(Utils.scala:538)
>     at
>
> kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:67)
>     at kafka.network.Receive$class.readCompletely(Transmission.scala:55)
>     at
>
> kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
>     at kafka.consumer.SimpleConsumer.getResponse(SimpleConsumer.scala:177)
>     at
> kafka.consumer.SimpleConsumer.liftedTree2$1(SimpleConsumer.scala:117)
>     at kafka.consumer.SimpleConsumer.multifetch(SimpleConsumer.scala:115)
>     at kafka.consumer.FetcherRunnable.run(FetcherRunnable.scala:60)
>
>
>
> 2013/3/19 Neha Narkhede <[EMAIL PROTECTED]>
>
> > Modify the log4j properties for senseidb and set kafka.consumer to INFO.
> > you can check the senseidb startup scripts on how they configure their
> > log4j.
> >
> > Thanks,
> > Neha
> >
> > On Tuesday, March 19, 2013, Yonghui Zhao wrote:
> >
> > > Hi Neha,
> > >
> > > How can I enable all kafka consumer log in senseidb?
> > > Btw: I am using kafka 0.7.2 java client.
> > >
> > > 2013/3/19 Neha Narkhede <[EMAIL PROTECTED] <javascript:;>>
> > >
> > > > The logs show that senseidb is prematurely closing the socket
> > connection
> > > to
> > > > the Kafka broker. I would enable atleast INFLO logging for Kafka in
> > > > Senseidb to see what the issue is.
> > > >
> > > > Thanks,
> > > > Neha
> > > >
> > > > On Monday, March 18, 2013, Yonghui Zhao wrote:
> > > >
> > > > > Thanks Neha,
> > > > >
> > > > > I use one kafka server with 4 partitions and 3 consumers(senseidb).
> > > > >
> > > > > Kafka server producer input rate is about 10k.
> > > > > And each consumer consuming rate is about 3k.
> > > > >
> > > > > I see this exceptions many times, kafka has this exception on each
> > > > > consumers, but I didn't find error log in consumer side,
> > > > consumer(senseidb)
> > > > > is alive all the time.
> > > > > Is it possible the exception is related with high input/output
> rate?
> > > > >
> > > > > And some times another exception(*Connection reset by peer*)
> > happened.
> > > > >
> > > > > [2013-03-18 21:18:29,107] ERROR Closing socket for
> > /10.2.201.203because
> > > > > of error (kafka.network.Processor)