Hi Peter,

The server is not restart. Anyway I will change log.dir, and will monitor
it again. Thanks for your help.

Best Regards,
Ben

Peter Davis <davi...@gmail.com>于2016年7月2日周六 下午12:15写道:

> Dear 黄杰斌:
>
> I am guessing your operating system is configured to delete your /tmp
> directory when you restart the server.
>
> You will need to change the "log.dir" property in your broker's
> server.properties file to someplace permanent.  Unfortunately, your data is
> lost unless you had a backup or had configured replication.
>
> log.dir The directory in which the log data is kept (supplemental for
> log.dirs property)        string  /tmp/kafka-logs         high
>
>
> Dear Community: why does log.dir default under /tmp?  It is unsafe as a
> default.
>
> -Peter
>
>
> > On Jun 30, 2016, at 11:19 PM, 黄杰斌 <jben.hu...@gmail.com> wrote:
> >
> > Hi All,
> >
> > Do you encounter below issue when using kafka_2.11-0.10.0.0?
> > All brokers are crash due to __consumer_offsets folder are deleted.
> > sample log:
> > [2016-06-30 12:46:32,579] FATAL [Replica Manager on Broker 2]: Halting
> due
> > to unrecoverable I/O error while handling produce request:
> > (kafka.server.ReplicaManager)
> > kafka.common.KafkaStorageException: I/O exception in append to log
> > '__consumer_offsets-32'
> >        at kafka.log.Log.append(Log.scala:329)
> >        at kafka.cluster.Partition$$anonfun$11.apply(Partition.scala:443)
> >        at kafka.cluster.Partition$$anonfun$11.apply(Partition.scala:429)
> >        at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:231)
> >        at kafka.utils.CoreUtils$.inReadLock(CoreUtils.scala:237)
> >        at
> > kafka.cluster.Partition.appendMessagesToLeader(Partition.scala:429)
> >        at
> >
> kafka.server.ReplicaManager$$anonfun$appendToLocalLog$2.apply(ReplicaManager.scala:406)
> >        at
> >
> kafka.server.ReplicaManager$$anonfun$appendToLocalLog$2.apply(ReplicaManager.scala:392)
> >        at
> >
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
> >        at
> >
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
> >        at scala.collection.immutable.Map$Map1.foreach(Map.scala:116)
> >        at
> > scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
> >        at scala.collection.AbstractTraversable.map(Traversable.scala:104)
> >        at
> > kafka.server.ReplicaManager.appendToLocalLog(ReplicaManager.scala:392)
> >        at
> > kafka.server.ReplicaManager.appendMessages(ReplicaManager.scala:328)
> >        at
> >
> kafka.coordinator.GroupMetadataManager.store(GroupMetadataManager.scala:232)
> >        at
> >
> kafka.coordinator.GroupCoordinator$$anonfun$handleCommitOffsets$9.apply(GroupCoordinator.scala:424)
> >        at
> >
> kafka.coordinator.GroupCoordinator$$anonfun$handleCommitOffsets$9.apply(GroupCoordinator.scala:424)
> >        at scala.Option.foreach(Option.scala:257)
> >        at
> >
> kafka.coordinator.GroupCoordinator.handleCommitOffsets(GroupCoordinator.scala:424)
> >        at
> > kafka.server.KafkaApis.handleOffsetCommitRequest(KafkaApis.scala:310)
> >        at kafka.server.KafkaApis.handle(KafkaApis.scala:84)
> >        at
> > kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:60)
> >        at java.lang.Thread.run(Thread.java:745)
> > Caused by: java.io.FileNotFoundException:
> > /tmp/kafka2-logs/__consumer_offsets-32/00000000000000000000.index (No
> such
> > file or directory)
> >        at java.io.RandomAccessFile.open0(Native Method)
> >        at java.io.RandomAccessFile.open(RandomAccessFile.java:316)
> >        at java.io.RandomAccessFile.<init>(RandomAccessFile.java:243)
> >        at
> > kafka.log.OffsetIndex$$anonfun$resize$1.apply(OffsetIndex.scala:286)
> >        at
> > kafka.log.OffsetIndex$$anonfun$resize$1.apply(OffsetIndex.scala:285)
> >        at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:231)
> >        at kafka.log.OffsetIndex.resize(OffsetIndex.scala:285)
> >        at
> >
> kafka.log.OffsetIndex$$anonfun$trimToValidSize$1.apply$mcV$sp(OffsetIndex.scala:274)
> >        at
> >
> kafka.log.OffsetIndex$$anonfun$trimToValidSize$1.apply(OffsetIndex.scala:274)
> >        at
> >
> kafka.log.OffsetIndex$$anonfun$trimToValidSize$1.apply(OffsetIndex.scala:274)
> >        at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:231)
> >        at kafka.log.OffsetIndex.trimToValidSize(OffsetIndex.scala:273)
> >        at kafka.log.Log.roll(Log.scala:655)
> >        at kafka.log.Log.maybeRoll(Log.scala:630)
> >        at kafka.log.Log.append(Log.scala:383)
> >        ... 23 more
> >
> > No one remove those folders, and topic __consumer_offsets is handled by
> > broker, no one can remove this topic.
> > Do you know why this happened? And how to avoid it?
> >
> > Best Regards,
> > Ben
>

Reply via email to