Hi Aravind,

Aravind Divakaran wrote:
>> Hi Aravind,
>>
>> Aravind Divakaran wrote:
>>> Hi Tao,
>>>
>>>> Hi Aravind,
>>>>
>>>> Aravind Divakaran wrote:
>>>>> Hi All,
>>>>>
>>>>> I have already sent one mail regarding the space issue i am facing
>>>>> with
>>>>> my
>>>>> ocfs filesystem. As mentioned in the below link it is an issue related
>>>>> to
>>>>> free space fragmentation.
>>>>>
>>>>> http://oss.oracle.com/bugzilla/show_bug.cgi?id=1189
>>>>>
>>>>> I have seen a patch for stealing extent allocation which was there is
>>>>> 2.6.34-rc1 kernel. So i compiled my new kernel and installed on my
>>>>> system.
>>>>>
>>>>> Below is my ocfs details on my system
>>>>>
>>>>> #modinfo ocfs2
>>>>>
>>>>> filename:       /lib/modules/2.6.34-rc1/kernel/fs/ocfs2/ocfs2.ko
>>>>> license:        GPL
>>>>> author:         Oracle
>>>>> version:        1.5.0
>>>>> description:    OCFS2 1.5.0
>>>>> srcversion:     A8B69947E8FF56D74858993
>>>>> depends:        jbd2,ocfs2_stackglue,quota_tree,ocfs2_nodemanager
>>>>> vermagic:       2.6.34-rc1 SMP mod_unload modversions
>>>>>
>>>>> This is my stat_sysdir.sh output
>>>>>
>>>>> http://pastebin.com/RZH9DkTk
>>>>>
>>>>> Can anyone help me how to resolve this, please as the problem occurs
>>>>> on
>>>>> production mail server with 3000 emailid.
>>>> I just checked your stat_sysdir output. It isn't caused by extent block
>>>> alloc actually. So the patch doesn't work for you. Yes, the problem you
>>>> meet is fragmentation issue, but the root cause is that inode_alloc
>>>> can't allocate any more inodes(a little different from 1189).
>>>>
>>>> I am now working on discontiguous block group. It will resolve your
>>>> issue I think. Hope it can be get into mainline in 2.6.35.
>>>>
>>>> Regards,
>>>> Tao
>>>>
>>> For my previous mail i got reply from you
>>>
>>> "Another way is that you can cp the file to another volume, remove it
>>> and
>>> then cp back. It should be contiguous enough."
>>>
>>> As mentioned in the 1189
>>>
>>> "However, reducing the slot count by 1 (to 4) may not be enough as it
>>> does
>>> not
>>> have much contiguous space. It may work. But reducing it by 2 will
>>> definitely work.
>>>
>>> Umount the volume on all nodes and run:
>>> # tunefs.ocfs2 -N 3 /dev/sda1
>>>
>>> Run fsck.ocfs2 for sanity checking."
>>>
>>> Will anyone of the above solution will temporary solve my problem.
>> Yes, it works. I just replied you in another e-mail.
>>
>> Regards,
>> Tao
>>
> I am running tunefs.ocfs2 on my 500gb harddisk which contain 215gb of
> data, in order to reduce the slots. I had used the below command.
> 
> tunefs.ocfs2  -N 3 /dev/mapper/store
> 
> Now almost 7hours is over still it didnt finished the execution. Below is
> the output i am getting.
> 
> node01:~# tunefs.ocfs2 -N 3 /dev/mapper/store
> tunefs.ocfs2 1.4.1
> 
> How much time it will take to reduce the slots. Whether it will be
> finished within 10hours. Can anyone help me.
It shouldn't cost so much time. I guess it get blocked in some case. So 
is this volume umounted in all the nodes? If yes, could you please 
strace it to see what's wrong?

Regards,
Tao

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

Reply via email to