Hauptseite > Publikationsdatenbank > Execution Performance Analysis of the ABySS Genome Sequence Assembler using Scalasca on the K computer > print |
001 | 279895 | ||
005 | 20250314084113.0 | ||
037 | _ | _ | |a FZJ-2015-07771 |
100 | 1 | _ | |a Kitayama, Itaru |0 P:(DE-HGF)0 |b 0 |
111 | 2 | _ | |a International Conference on Parallel Computing |g ParCo |c Edinburgh |d 2015-09-01 - 2015-09-04 |w Scotland |
245 | _ | _ | |a Execution Performance Analysis of the ABySS Genome Sequence Assembler using Scalasca on the K computer |
260 | _ | _ | |c 2015 |
336 | 7 | _ | |a Conference Presentation |b conf |m conf |0 PUB:(DE-HGF)6 |s 1450443653_24783 |2 PUB:(DE-HGF) |x After Call |
336 | 7 | _ | |a Conference Paper |0 33 |2 EndNote |
336 | 7 | _ | |a Other |2 DataCite |
336 | 7 | _ | |a LECTURE_SPEECH |2 ORCID |
336 | 7 | _ | |a conferenceObject |2 DRIVER |
336 | 7 | _ | |a INPROCEEDINGS |2 BibTeX |
500 | _ | _ | |a pdf darf NICHT open access sein |
520 | _ | _ | |a Performance analysis of the ABySS genome sequence assembler (ABYSS-P) executing on the K computer with up to 8192 compute nodes is described which identified issues that limited scalability to less than 1024 compute nodes and required prohibitive message buffer memory with 16384 or more compute nodes. The open-source Scalasca toolset was employed to analyse executions, revealing the impact of massive amounts of MPI point-to-point communication used particularly for master/worker process coordination, and inefficient parallel file operations that manifest as waiting time at later MPI collective synchronisations and communications. Initial remediation via use of collective communication operations and alternate strategies for parallel file handling show large performance and scalability improvements, with partial executions validated on the full 82,944 compute nodes of the K computer. |
536 | _ | _ | |a 511 - Computational Science and Mathematical Methods (POF3-511) |0 G:(DE-HGF)POF3-511 |c POF3-511 |f POF III |x 0 |
536 | _ | _ | |0 G:(DE-Juel-1)ATMLPP |a ATMLPP - ATML Parallel Performance (ATMLPP) |c ATMLPP |x 1 |
700 | 1 | _ | |a Wylie, Brian J. N. |0 P:(DE-Juel1)132302 |b 1 |e Corresponding author |u fzj |
700 | 1 | _ | |a Maeda, Toshiyuki |0 P:(DE-HGF)0 |b 2 |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.pdf |y Restricted |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.gif?subformat=icon |x icon |y Restricted |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.jpg?subformat=icon-1440 |x icon-1440 |y Restricted |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.jpg?subformat=icon-180 |x icon-180 |y Restricted |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.jpg?subformat=icon-640 |x icon-640 |y Restricted |
856 | 4 | _ | |u https://juser.fz-juelich.de/record/279895/files/ParCo2015_wylie.pdf?subformat=pdfa |x pdfa |y Restricted |
909 | C | O | |o oai:juser.fz-juelich.de:279895 |p VDB |
910 | 1 | _ | |a Forschungszentrum Jülich GmbH |0 I:(DE-588b)5008462-8 |k FZJ |b 1 |6 P:(DE-Juel1)132302 |
913 | 1 | _ | |a DE-HGF |b Key Technologies |1 G:(DE-HGF)POF3-510 |0 G:(DE-HGF)POF3-511 |2 G:(DE-HGF)POF3-500 |v Computational Science and Mathematical Methods |x 0 |4 G:(DE-HGF)POF |3 G:(DE-HGF)POF3 |l Supercomputing & Big Data |
914 | 1 | _ | |y 2015 |
920 | _ | _ | |l yes |
920 | 1 | _ | |0 I:(DE-Juel1)JSC-20090406 |k JSC |l Jülich Supercomputing Center |x 0 |
980 | _ | _ | |a conf |
980 | _ | _ | |a VDB |
980 | _ | _ | |a I:(DE-Juel1)JSC-20090406 |
980 | _ | _ | |a UNRESTRICTED |
Library | Collection | CLSMajor | CLSMinor | Language | Author |
---|