thanks for the bug report. the development team will take a look at the changes to the 2.5.30 kernel and /proc/stat and be sure to update our code.
-- matt Sep 24, Benoit des Ligneris wrote forth saying... > Hello, > > We tried to use ganglia on diskless nodes with system in a ramdisk. We > used a 2.5.30 kernel and gmond is dying after having send one sample of > it's metrics. > > Here is the log : > ================= > > set_metric_value() got metric key 20 > set_metric_value() exec'd mem_free_func (20) > mcast_value() mcasting mem_free value > encoded 8 XDR bytes > XDR data successfully sent > set_metric_value() got metric key 21 > set_metric_value() exec'd mem_shared_func (21) > mcast_value() mcasting mem_shared value > encoded 8 XDR bytes > XDR data successfully sent > set_metric_value() got metric key 22 > set_metric_value() exec'd mem_buffers_func (22) > Segmentation fault > > > A strace does not give additional information (I can send the strace log if > requested) > and all is identical (libraries, ...) to the "master node" > where gmond runs fine excepts : > - kernel 2.4.18 > - /usr directory is read-only on the nodes. > > We tried ganglia 2.4.1 and 2.5. > > I noticed that /proc/stat is different for the two kernels and so I > wonder if it can cause gmond to crash "badly" : > > Node (kernel 2.5.30) > > ************************* oscar_cluster ************************* > processing node node01.tour1 > --------- node01.tour1--------- > cpu 152060 6 12405 8089102 > cpu0 152060 6 12405 8089102 > page 7446 18246 > swap 0 0 > intr 84250671 82535754 2 0 0 0 0 0 0 0 0 0 1714915 0 0 0 0 > disk_io: > pageallocs 6662731 > pagefrees 6632641 > pageactiv 7247 > pagedeact 0 > pagefault 15571178 > majorfault 833 > pagescan 0 > pagesteal 0 > pageoutrun 0 > allocstall 0 > ctxt 2009358 > btime 1032817204 > processes 119263 > > Master (2.4.18) > > [EMAIL PROTECTED] root]# cat /proc/stat > cpu 1701148 219838 666892 181036471 > cpu0 1701148 219838 666892 181036471 > page 9091489 51558917 > swap 40293 49304 > intr 262331317 183624349 28077 0 15878 3 21714 2 0 1 0 2472287 70258953 > 160026 0 5750027 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 > disk_io: (3,0):(5761797,811852,18119716,4949945,102072642) > ctxt 802893759 > btime 1031063501 > processes 242532 > >