Poster (After Call) FZJ-2022-03599

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
OpenGPT-X - Training Large Language Models on HPC Systems

 ;  ;  ;  ;  ;

2022

14th JLESC Workshop, Urbana-ChampaignUrbana-Champaign, USA, 28 Sep 2022 - 30 Sep 20222022-09-282022-09-30

Please use a persistent id in citations:

Abstract: Artificial neural networks represent an HPC workload with increasing importance. In particular the field of Natural Language Processing (NLP) has been undergoing a revolution in recent years. The training of ever larger language models, such as GPT-3, demands large HPC resources and has the potential to greatly impact everyday technology. The OpenGPT-X project was established in 2022 and aims to not leave this field to large tech companies but to provide an open, publicly funded alternative based on European values. The Jülich Supercomputing Centre is a consortium partner providing HPC infrastructure for the pre-training of the models. We research the optimization potential in the training process for example by using novel accelerator architectures.


Contributing Institute(s):
  1. Jülich Supercomputing Center (JSC)
Research Program(s):
  1. 5112 - Cross-Domain Algorithms, Tools, Methods Labs (ATMLs) and Research Groups (POF4-511) (POF4-511)

Appears in the scientific report 2022
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Document types > Presentations > Poster
Workflow collections > Public records
Institute Collections > JSC
JuOSC (Juelich Open Science Collection)
Publications database
Open Access

 Record created 2022-10-06, last modified 2023-07-12


OpenAccess:
Download fulltext PDF
External link:
Download fulltextFulltext by OpenAccess repository
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)