000845503 001__ 845503
000845503 005__ 20250317091731.0
000845503 0247_ $$2doi$$a10.14529/jsfi180104
000845503 0247_ $$2ISSN$$a2313-8734
000845503 0247_ $$2ISSN$$a2409-6008
000845503 0247_ $$2Handle$$a2128/18391
000845503 037__ $$aFZJ-2018-02737
000845503 041__ $$aEnglish
000845503 082__ $$a004
000845503 1001_ $$0P:(DE-Juel1)143606$$aBrömmel, Dirk$$b0$$ufzj
000845503 245__ $$aThe High-Q Club: Experience with Extreme-scaling Application Codes
000845503 260__ $$aChelyabinsk$$bSouth Ural State University$$c2018
000845503 3367_ $$2DRIVER$$aarticle
000845503 3367_ $$2DataCite$$aOutput Types/Journal article
000845503 3367_ $$0PUB:(DE-HGF)16$$2PUB:(DE-HGF)$$aJournal Article$$bjournal$$mjournal$$s1525352015_15477
000845503 3367_ $$2BibTeX$$aARTICLE
000845503 3367_ $$2ORCID$$aJOURNAL_ARTICLE
000845503 3367_ $$00$$2EndNote$$aJournal Article
000845503 520__ $$aJülich Supercomputing Centre (JSC) started running (extreme) scaling workshops with its first IBM Blue Gene supercomputer, finally spanning three generations each seeing an increase in the number of cores and available threads. Over the years, this workshop series attracted numerous international code teams and resulted in many applications capable of running on all available cores of each system.This article reviews some of the knowledge gained with running and tuning highly-scalable applications, focussing on JUQUEEN, the IBM Blue Gene/Q at JSC. The ability to execute successfully on all 458752 cores with up to 1.8 million processes or threads may qualify codes for the High-Q Club, which serves as a showcase for diverse codes scaling to the entire 28 racks, effectively defining a collection of the highest scaling codes on JUQUEEN. The intention was to encourage other developers to invest in tuning and scaling their codes while identifying the necessary key aspects for that goal.As this era closes, it is timely to compare the characteristics of the 32 High-Q Club member codes, considering their strong and/or weak scaling, exploitation of hardware threading, and whether/how intra-node multi-threading is employed combined with message-passing. We also identify the obstacles for scaling such as inefficient use of limited compute node memory and file I/O as key governing factors. Overall, the analysis provides guidance as to how applications may (need to) be designed in future to exploit expected exa-scale computer systems.
000845503 536__ $$0G:(DE-HGF)POF3-511$$a511 - Computational Science and Mathematical Methods (POF3-511)$$cPOF3-511$$fPOF III$$x0
000845503 536__ $$0G:(DE-HGF)POF3-513$$a513 - Supercomputer Facility (POF3-513)$$cPOF3-513$$fPOF III$$x1
000845503 536__ $$0G:(DE-Juel-1)ATMLPP$$aATMLPP - ATML Parallel Performance (ATMLPP)$$cATMLPP$$x2
000845503 536__ $$0G:(DE-Juel-1)ATMLAO$$aATMLAO - ATML Application Optimization and User Service Tools (ATMLAO)$$cATMLAO$$x3
000845503 588__ $$aDataset connected to CrossRef
000845503 7001_ $$0P:(DE-Juel1)132108$$aFrings, Wolfgang$$b1$$ufzj
000845503 7001_ $$0P:(DE-Juel1)132302$$aWylie, Brian J. N.$$b2$$ufzj
000845503 7001_ $$0P:(DE-Juel1)132199$$aMohr, Bernd$$b3$$ufzj
000845503 7001_ $$0P:(DE-Juel1)132115$$aGibbon, Paul$$b4$$ufzj
000845503 7001_ $$0P:(DE-Juel1)132179$$aLippert, Thomas$$b5$$ufzj
000845503 773__ $$0PERI:(DE-600)2809718-X$$a10.14529/jsfi180104$$gVol. 5, no. 1$$n1$$p59-78$$tSupercomputing frontiers and innovations$$v5$$x2313-8734$$y2018
000845503 8564_ $$uhttps://juser.fz-juelich.de/record/845503/files/HiQ.pdf$$yOpenAccess
000845503 8564_ $$uhttps://juser.fz-juelich.de/record/845503/files/HiQ.gif?subformat=icon$$xicon$$yOpenAccess
000845503 8564_ $$uhttps://juser.fz-juelich.de/record/845503/files/HiQ.jpg?subformat=icon-180$$xicon-180$$yOpenAccess
000845503 8564_ $$uhttps://juser.fz-juelich.de/record/845503/files/HiQ.jpg?subformat=icon-700$$xicon-700$$yOpenAccess
000845503 909CO $$ooai:juser.fz-juelich.de:845503$$pdnbdelivery$$pdriver$$pVDB$$popen_access$$popenaire
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)143606$$aForschungszentrum Jülich$$b0$$kFZJ
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)132108$$aForschungszentrum Jülich$$b1$$kFZJ
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)132302$$aForschungszentrum Jülich$$b2$$kFZJ
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)132199$$aForschungszentrum Jülich$$b3$$kFZJ
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)132115$$aForschungszentrum Jülich$$b4$$kFZJ
000845503 9101_ $$0I:(DE-588b)5008462-8$$6P:(DE-Juel1)132179$$aForschungszentrum Jülich$$b5$$kFZJ
000845503 9131_ $$0G:(DE-HGF)POF3-511$$1G:(DE-HGF)POF3-510$$2G:(DE-HGF)POF3-500$$3G:(DE-HGF)POF3$$4G:(DE-HGF)POF$$aDE-HGF$$bKey Technologies$$lSupercomputing & Big Data$$vComputational Science and Mathematical Methods$$x0
000845503 9131_ $$0G:(DE-HGF)POF3-513$$1G:(DE-HGF)POF3-510$$2G:(DE-HGF)POF3-500$$3G:(DE-HGF)POF3$$4G:(DE-HGF)POF$$aDE-HGF$$bKey Technologies$$lSupercomputing & Big Data$$vSupercomputer Facility$$x1
000845503 9141_ $$y2018
000845503 915__ $$0StatID:(DE-HGF)0510$$2StatID$$aOpenAccess
000845503 915__ $$0StatID:(DE-HGF)0040$$2StatID$$aPeer Review unknown
000845503 920__ $$lyes
000845503 9201_ $$0I:(DE-Juel1)JSC-20090406$$kJSC$$lJülich Supercomputing Center$$x0
000845503 980__ $$ajournal
000845503 980__ $$aVDB
000845503 980__ $$aUNRESTRICTED
000845503 980__ $$aI:(DE-Juel1)JSC-20090406
000845503 9801_ $$aFullTexts