[gpfsug-discuss] Critical Hang issues with GPFS 5.0. Downgrading from GPFS 5.0.0-2 to GPFS 4.2.3.2

valleru at cbio.mskcc.org valleru at cbio.mskcc.org
Fri Nov 2 16:31:12 GMT 2018


Also - You could just upgrade one of the clients to this version, and test to see if the hang still occurs.
You do not have to upgrade the NSD servers, to test.

Regards,
Lohit

On Nov 2, 2018, 12:29 PM -0400, valleru at cbio.mskcc.org, wrote:
> Yes,
>
> We have upgraded to 5.0.1-0.5, which has the patch for the issue.
> The related IBM case number was : TS001010674
>
> Regards,
> Lohit
>
> On Nov 2, 2018, 12:27 PM -0400, Mazurkova, Svetlana/Information Systems <sveta at cbio.mskcc.org>, wrote:
> > Hi Damir,
> >
> > It was related to specific user jobs and mmap (?). We opened PMR with IBM and have patch from IBM, since than we don’t see issue.
> >
> > Regards,
> >
> > Sveta.
> >
> > > On Nov 2, 2018, at 11:55 AM, Damir Krstic <damir.krstic at gmail.com> wrote:
> > >
> > > Hi,
> > >
> > > Did you ever figure out the root cause of the issue? We have recently (end of the June) upgraded our storage to: gpfs.base-5.0.0-1.1.3.ppc64
> > >
> > > In the last few weeks we have seen an increasing number of ps hangs across compute and login nodes on our cluster. The filesystem version (of all filesystems on our cluster) is:
> > >  -V                 15.01 (4.2.0.0)          File system version
> > >
> > > I am just wondering if anyone has seen this type of issue since you first reported it and if there is a known fix for it.
> > >
> > > Damir
> > >
> > > > On Tue, May 22, 2018 at 10:43 AM <valleru at cbio.mskcc.org> wrote:
> > > > > Hello All,
> > > > >
> > > > > We have recently upgraded from GPFS 4.2.3.2 to GPFS 5.0.0-2 about a month ago. We have not yet converted the 4.2.2.2 filesystem version to 5. ( That is we have not run the mmchconfig release=LATEST command)
> > > > > Right after the upgrade, we are seeing many “ps hangs" across the cluster. All the “ps hangs” happen when jobs run related to a Java process or many Java threads (example: GATK )
> > > > > The hangs are pretty random, and have no particular pattern except that we know that it is related to just Java or some jobs reading from directories with about 600000 files.
> > > > >
> > > > > I have raised an IBM critical service request about a month ago related to this - PMR: 24090,L6Q,000.
> > > > > However, According to the ticket  - they seemed to feel that it might not be related to GPFS.
> > > > > Although, we are sure that these hangs started to appear only after we upgraded GPFS to GPFS 5.0.0.2 from 4.2.3.2.
> > > > >
> > > > > One of the other reasons we are not able to prove that it is GPFS is because, we are unable to capture any logs/traces from GPFS once the hang happens.
> > > > > Even GPFS trace commands hang, once “ps hangs” and thus it is getting difficult to get any dumps from GPFS.
> > > > >
> > > > > Also  - According to the IBM ticket, they seemed to have a seen a “ps hang" issue and we have to run  mmchconfig release=LATEST command, and that will resolve the issue.
> > > > > However we are not comfortable making the permanent change to Filesystem version 5. and since we don’t see any near solution to these hangs - we are thinking of downgrading to GPFS 4.2.3.2 or the previous state that we know the cluster was stable.
> > > > >
> > > > > Can downgrading GPFS take us back to exactly the previous GPFS config state?
> > > > > With respect to downgrading from 5 to 4.2.3.2 -> is it just that i reinstall all rpms to a previous version? or is there anything else that i need to make sure with respect to GPFS configuration?
> > > > > Because i think that GPFS 5.0 might have updated internal default GPFS configuration parameters , and i am not sure if downgrading GPFS will change them back to what they were in GPFS 4.2.3.2
> > > > >
> > > > > Our previous state:
> > > > >
> > > > > 2 Storage clusters - 4.2.3.2
> > > > > 1 Compute cluster - 4.2.3.2  ( remote mounts the above 2 storage clusters )
> > > > >
> > > > > Our current state:
> > > > >
> > > > > 2 Storage clusters - 5.0.0.2 ( filesystem version - 4.2.2.2)
> > > > > 1 Compute cluster - 5.0.0.2
> > > > >
> > > > > Do i need to downgrade all the clusters to go to the previous state ? or is it ok if we just downgrade the compute cluster to previous version?
> > > > >
> > > > > Any advice on the best steps forward, would greatly help.
> > > > >
> > > > > Thanks,
> > > > >
> > > > > Lohit
> > > > > _______________________________________________
> > > > > gpfsug-discuss mailing list
> > > > > gpfsug-discuss at spectrumscale.org
> > > > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
> > > _______________________________________________
> > > gpfsug-discuss mailing list
> > > gpfsug-discuss at spectrumscale.org
> > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
> >
> > _______________________________________________
> > gpfsug-discuss mailing list
> > gpfsug-discuss at spectrumscale.org
> > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20181102/6dcd7b3f/attachment.htm>


More information about the gpfsug-discuss mailing list