Hi, All

We fix the problem, I like to share the what the problem is in case someone
come across the similar issues. We add the data drive for each node
/dev/sdb1 , but specify the wrong path in server.properties, which means
the data was written into the wrong drive /dev/sda2, quickly eat up all the
space in sda2, now we change the path. The sdb1 has 15Tb, which allows us
to store data for a while and will be deleted in 1/2 weeks as config
mentioned.

But I am kinda curious about David's comments,  "... after having tuned
retention bytes or retention (time?) incorrectly. .."  How do you guys set
log.retention.bytes?  I set log.retention.hours=336 (2 weeks), and should I
set log.retention.bytes as default -1 or some other amount?

thanks

AL

On Tue, Jan 6, 2015 at 12:43 PM, Sa Li <sal...@gmail.com> wrote:

> Thanks the reply, the disk is not full:
>
> root@exemplary-birds:~# df -h
> Filesystem      Size  Used Avail Use% Mounted on
> /dev/sda2       133G  3.4G  123G   3% /
> none            4.0K     0  4.0K   0% /sys/fs/cgroup
> udev             32G  4.0K   32G   1% /dev
> tmpfs           6.3G  764K  6.3G   1% /run
> none            5.0M     0  5.0M   0% /run/lock
> none             32G     0   32G   0% /run/shm
> none            100M     0  100M   0% /run/user
> /dev/sdb1        14T   15G   14T   1% /srv
>
> Neither the memory
>
> root@exemplary-birds:~# free
>              total       used       free     shared    buffers     cached
> Mem:      65963372    9698380   56264992        776     170668    7863812
> -/+ buffers/cache:    1663900   64299472
> Swap:       997372          0     997372
>
> thanks
>
>
> On Tue, Jan 6, 2015 at 12:10 PM, David Birdsong <david.birds...@gmail.com>
> wrote:
>
>> I'm keen to hear about how to work one's way out of a filled partition
>> since I've run into this many times after having tuned retention bytes or
>> retention (time?) incorrectly. The proper path to resolving this isn't
>> obvious based on my many harried searches through documentation.
>>
>> I often end up stopping the particular broker, picking an unlucky
>> topic/partition, deleting, modifying the any topics that consumed too much
>> space by lowering their retention bytes, and restarting.
>>
>> On Tue, Jan 6, 2015 at 12:02 PM, Sa Li <sal...@gmail.com> wrote:
>>
>> > Continue this issue, when I restart the server, like
>> > bin/kafka-server-start.sh config/server.properties
>> >
>> > it will fails to start the server, like
>> >
>> > [2015-01-06 20:00:55,441] FATAL Fatal error during KafkaServerStable
>> > startup. Prepare to shutdown (kafka.server.KafkaServerStartable)
>> > java.lang.InternalError: a fault occurred in a recent unsafe memory
>> access
>> > operation in compiled Java code
>> >         at java.nio.HeapByteBuffer.<init>(HeapByteBuffer.java:57)
>> >         at java.nio.ByteBuffer.allocate(ByteBuffer.java:331)
>> >         at
>> > kafka.log.FileMessageSet$$anon$1.makeNext(FileMessageSet.scala:188)
>> >         at
>> > kafka.log.FileMessageSet$$anon$1.makeNext(FileMessageSet.scala:165)
>> >         at
>> > kafka.utils.IteratorTemplate.maybeComputeNext(IteratorTemplate.scala:66)
>> >         at
>> kafka.utils.IteratorTemplate.hasNext(IteratorTemplate.scala:58)
>> >         at kafka.log.LogSegment.recover(LogSegment.scala:165)
>> >         at kafka.log.Log.recoverLog(Log.scala:179)
>> >         at kafka.log.Log.loadSegments(Log.scala:155)
>> >         at kafka.log.Log.<init>(Log.scala:64)
>> >         at
>> >
>> >
>> kafka.log.LogManager$$anonfun$loadLogs$1$$anonfun$apply$4.apply(LogManager.scala:118)
>> >         at
>> >
>> >
>> kafka.log.LogManager$$anonfun$loadLogs$1$$anonfun$apply$4.apply(LogManager.scala:113)
>> >         at
>> >
>> >
>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>> >         at
>> > scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:105)
>> >         at
>> > kafka.log.LogManager$$anonfun$loadLogs$1.apply(LogManager.scala:113)
>> >         at
>> > kafka.log.LogManager$$anonfun$loadLogs$1.apply(LogManager.scala:105)
>> >         at
>> >
>> >
>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>> >         at
>> > scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:34)
>> >         at kafka.log.LogManager.loadLogs(LogManager.scala:105)
>> >         at kafka.log.LogManager.<init>(LogManager.scala:57)
>> >         at
>> kafka.server.KafkaServer.createLogManager(KafkaServer.scala:275)
>> >         at kafka.server.KafkaServer.startup(KafkaServer.scala:72)
>> >         at
>> > kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:34)
>> >         at kafka.Kafka$.main(Kafka.scala:46)
>> >         at kafka.Kafka.main(Kafka.scala)
>> > [2015-01-06 20:00:55,443] INFO [Kafka Server 100], shutting down
>> > (kafka.server.KafkaServer)
>> > [2015-01-06 20:00:55,444] INFO Terminate ZkClient event thread.
>> > (org.I0Itec.zkclient.ZkEventThread)
>> > [2015-01-06 20:00:55,446] INFO Session: 0x684a5ed9da3a1a0f closed
>> > (org.apache.zookeeper.ZooKeeper)
>> > [2015-01-06 20:00:55,446] INFO EventThread shut down
>> > (org.apache.zookeeper.ClientCnxn)
>> > [2015-01-06 20:00:55,447] INFO [Kafka Server 100], shut down completed
>> > (kafka.server.KafkaServer)
>> > [2015-01-06 20:00:55,447] INFO [Kafka Server 100], shutting down
>> > (kafka.server.KafkaServer)
>> >
>> > Any ideas
>> >
>> > On Tue, Jan 6, 2015 at 12:00 PM, Sa Li <sal...@gmail.com> wrote:
>> >
>> > > the complete error message:
>> > >
>> > > -su: cannot create temp file for here-document: No space left on
>> device
>> > > OpenJDK 64-Bit Server VM warning: Insufficient space for shared memory
>> > > file:
>> > >    /tmp/hsperfdata_root/19721
>> > > Try using the -Djava.io.tmpdir= option to select an alternate temp
>> > > location.
>> > > [2015-01-06 19:50:49,244] FATAL  (kafka.Kafka$)
>> > > java.io.FileNotFoundException: conf (No such file or directory)
>> > >         at java.io.FileInputStream.open(Native Method)
>> > >         at java.io.FileInputStream.<init>(FileInputStream.java:146)
>> > >         at java.io.FileInputStream.<init>(FileInputStream.java:101)
>> > >         at kafka.utils.Utils$.loadProps(Utils.scala:144)
>> > >         at kafka.Kafka$.main(Kafka.scala:34)
>> > >         at kafka.Kafka.main(Kafka.scala)
>> > >
>> > > On Tue, Jan 6, 2015 at 11:58 AM, Sa Li <sal...@gmail.com> wrote:
>> > >
>> > >>
>> > >> Hi, All
>> > >>
>> > >> I am doing performance test on our new kafka production server, but
>> > after
>> > >> sending some messages (even faked message by using
>> > bin/kafka-run-class.sh
>> > >> org.apache.kafka.clients.tools.ProducerPerformance), it comes out the
>> > error
>> > >> of connection, and shut down the brokers, after that, I see such
>> errors,
>> > >>
>> > >> conf-su: cannot create temp file for here-document: No space left on
>> > >> device
>> > >>
>> > >> How can I fix it, I am concerning that will happen when we start to
>> > >> publish real messages in kafka, and should I create some cron to
>> > regularly
>> > >> clean certain directories?
>> > >>
>> > >> thanks
>> > >>
>> > >> --
>> > >>
>> > >> Alec Li
>> > >>
>> > >
>> > >
>> > >
>> > > --
>> > >
>> > > Alec Li
>> > >
>> >
>> >
>> >
>> > --
>> >
>> > Alec Li
>> >
>>
>
>
>
> --
>
> Alec Li
>



-- 

Alec Li

Reply via email to