Yaron, doesn't "-1" make each of these disk an independent failure group?
>From 'man mmcrnsd': "The default is -1, which indicates this disk has no point of failure in common with any other disk." -jf man. 9. jan. 2017 kl. 21.54 skrev Yaron Daniel <y...@il.ibm.com>: > Hi > > So - do u able to have GPFS replication > > for the MD Failure Groups ? > > I can see that u have 3 Failure Groups > > for Data -1, 2012,2034 , how many Storage Subsystems you have ? > > > > > Regards > > > > ------------------------------ > > > > > > *YaronDaniel* 94 > > Em Ha'Moshavot Rd > > > *Server,* > > *Storageand Data Services* > <https://w3-03.ibm.com/services/isd/secure/client.wss/Somt?eventType=getHomePage&somtId=115>*- > Team Leader* > > Petach > > Tiqva, 49527 > > > *GlobalTechnology Services* Israel > Phone: +972-3-916-5672 > Fax: +972-3-916-5672 > > > Mobile: +972-52-8395593 > > > e-mail: y...@il.ibm.com > > > > > *IBMIsrael* <http://www.ibm.com/il/he/> > > > > > > > > > > From: > > "J. Eric Wonderley" > > <eric.wonder...@vt.edu> > > > To: > > gpfsug main discussion > > list <gpfsug-discuss@spectrumscale.org> > > Date: > > 01/09/2017 10:48 PM > Subject: > > Re: [gpfsug-discuss] > > replication and no failure groups > Sent by: > > gpfsug-discuss-boun...@spectrumscale.org > ------------------------------ > > > > Hi Yaron: > > This is the filesystem: > > [root@cl005 net]# mmlsdisk work > disk driver > > sector failure holds holds > > storage > name type > > size group metadata data status > > availability pool > ------------ -------- ------ ----------- -------- ----- ------------- > ------------ > > ------------ > nsd_a_7 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_7 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_7 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_7 nsd > > 512 -1 No > > Yes ready up > > system > nsd_a_8 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_8 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_8 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_8 nsd > > 512 -1 No > > Yes ready up > > system > nsd_a_9 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_9 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_9 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_9 nsd > > 512 -1 No > > Yes ready up > > system > nsd_a_10 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_10 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_10 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_10 nsd > > 512 -1 No > > Yes ready up > > system > nsd_a_11 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_11 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_11 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_11 nsd > > 512 -1 No > > Yes ready up > > system > nsd_a_12 nsd > > 512 -1 No > > Yes ready up > > system > nsd_b_12 nsd > > 512 -1 No > > Yes ready up > > system > nsd_c_12 nsd > > 512 -1 No > > Yes ready up > > system > nsd_d_12 nsd > > 512 -1 No > > Yes ready up > > system > work_md_pf1_1 nsd 512 > > 200 Yes No ready > > up system > > > jbf1z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf2z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf3z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf4z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf5z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf6z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf7z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf8z1 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf1z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf2z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf3z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf4z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf5z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf6z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf7z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf8z2 nsd > > 4096 2012 No > > Yes ready up > > sas_ssd4T > jbf1z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf2z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf3z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf4z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf5z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf6z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf7z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf8z3 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf1z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf2z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf3z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf4z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf5z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf6z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf7z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > jbf8z4 nsd > > 4096 2034 No > > Yes ready up > > sas_ssd4T > work_md_pf1_2 nsd 512 > > 200 Yes No ready > > up system > > > work_md_pf1_3 nsd 512 > > 200 Yes No ready > > up system > > > work_md_pf1_4 nsd 512 > > 200 Yes No ready > > up system > > > work_md_pf2_5 nsd 512 > > 199 Yes No ready > > up system > > > work_md_pf2_6 nsd 512 > > 199 Yes No ready > > up system > > > work_md_pf2_7 nsd 512 > > 199 Yes No ready > > up system > > > work_md_pf2_8 nsd 512 > > 199 Yes No ready > > up system > > > [root@cl005 net]# mmlsfs work -R -r -M -m -K > flag > > value > > description > ------------------- ------------------------ > ----------------------------------- > -R > > 2 > > Maximum number of data replicas > -r > > 2 > > Default number of data replicas > -M > > 2 > > Maximum number of metadata replicas > -m > > 2 > > Default number of metadata replicas > -K > > whenpossible > > Strict replica allocation option > > > On Mon, Jan 9, 2017 at 3:34 PM, Yaron Daniel <*y...@il.ibm.com* > <y...@il.ibm.com>> > > wrote: > Hi > > 1) Yes in case u have only 1 Failure group - replication will not work. > > 2) Do you have 2 Storage Systems ? When using GPFS replication write > > stay the same - but read can be double - since it read from 2 Storage > systems > > Hope this help - what do you try to achive , can you share your env setup > > ? > > > Regards > > > > ------------------------------ > > > > > > *YaronDaniel* 94 > > Em Ha'Moshavot Rd > > > *Server,* > > *Storageand Data Services* > <https://w3-03.ibm.com/services/isd/secure/client.wss/Somt?eventType=getHomePage&somtId=115> > > *-Team Leader* Petach > > Tiqva, 49527 > > > *GlobalTechnology Services* Israel > Phone: *+972-3-916-5672* <+972%203-916-5672> > Fax: *+972-3-916-5672* <+972%203-916-5672> > > > Mobile: *+972-52-8395593* <+972%2052-839-5593> > > > e-mail: *y...@il.ibm.com* <y...@il.ibm.com> > > > > > *IBMIsrael* <http://www.ibm.com/il/he/> > > > > > > > > > > From: Brian > > Marshall <*mimar...@vt.edu* <mimar...@vt.edu>> > To: gpfsug > > main discussion list <*gpfsug-discuss@spectrumscale.org* > <gpfsug-discuss@spectrumscale.org>> > Date: 01/09/2017 > > 10:17 PM > Subject: [gpfsug-discuss] > > replication and no failure groups > Sent by: *gpfsug-discuss-boun...@spectrumscale.org* > <gpfsug-discuss-boun...@spectrumscale.org> > > ------------------------------ > > > > > All, > > If I have a filesystem with replication set to 2 and 1 failure group: > > 1) I assume replication won't actually happen, correct? > > 2) Will this impact performance i.e cut write performance in half even > > though it really only keeps 1 copy? > > End goal - I would like a single storage pool within the filesystem to > > be replicated without affecting the performance of all other pools(which > > only have a single failure group) > > Thanks, > Brian Marshall > VT - ARC_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at *spectrumscale.org* <http://spectrumscale.org/> > *http://gpfsug.org/mailman/listinfo/gpfsug-discuss* > <http://gpfsug.org/mailman/listinfo/gpfsug-discuss> > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at *spectrumscale.org* <http://spectrumscale.org/> > *http://gpfsug.org/mailman/listinfo/gpfsug-discuss* > <http://gpfsug.org/mailman/listinfo/gpfsug-discuss> > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > > > _______________________________________________ > > gpfsug-discuss mailing list > > gpfsug-discuss at spectrumscale.org > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss