Advances in High Performance Computing and Scalable Software

A special issue of Information (ISSN 2078-2489). This special issue belongs to the section "Information Applications".

Deadline for manuscript submissions: 30 September 2024 | Viewed by 6178

Special Issue Editors


E-Mail Website
Guest Editor
Department of Computer Science, University at Albany, Albany, NY 12222, USA
Interests: high performance computing; arrays; tensors; scalable software; optimizations

E-Mail Website
Guest Editor
1. Department of Computer Science, National University of Singapore, Singapore City, Singapore
2. Vq Research, Inc., Mountain View, CA 94043, USA
Interests: parallel computing; numerical analysis

Special Issue Information

Dear Colleagues,

HPC and scalable software pose numerous challenges to developers: achieving scalable performance, numerical accuracy and bitwise reproducibility, and ease of programming diverse architectures through tools (OpenMP, OpenACC, etc.) as well as advanced compiler techniques to allow application programmers to exploit current and future hardware designs without exposing them to machine-specific details. New fields like performance engineering are emerging to address some of these issues. However, without strong mathematical foundations and a commitment to codesign all components of software and hardware systems, such problems will increase and expand in complexity. Is it time to reformulate operating systems and memory management? Can existing languages still work in an era where massive HPC and AI data structures like arrays (tensors) must map automatically to a vast range of distributed memory designs? How can numerical accuracy and bitwise reproducibility be guaranteed on systems where the arithmetic and language standards do not specify bitwise-reproducible rounding? This Special Issue will address all those who are using mathematical foundations to seek portable and scalable ways to optimize hardware utilization without sacrificing programmer productivity, who use co-design to guarantee performance and accuracy, and who seek verifications of design. We can prove semantic designs, but can we prove operational designs? Can we prove that the outcome of two compilers on a single machine are equivalent? Can we predict performance automatically (using source code and a sufficient description of target hardware) to reduce the need for “ninja programmers”? What issues should be studied that have been avoided to solve these problems? This Special Issues encourages visionary scientists to submit papers in the areas of mathematical foundations to address all topics above, as well as the following:

(1) Numerical accuracy and bitwise reproducibility across machines and languages;

(2) Verification of semantic and operational designs;

(3) Domain-specific machines, operating systems, and languages;

(4) Software tools to automate scalable performance on HPC machines without sacrificing reproducibility.

Prof. Dr. Lenore Mullin
Prof. Dr. John L. Gustafson
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Information is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Published Papers (3 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

19 pages, 1277 KiB  
Article
Top-Down Models across CPU Architectures: Applicability and Comparison in a High-Performance Computing Environment
by Fabio Banchelli, Marta Garcia-Gasulla and Filippo Mantovani
Information 2023, 14(10), 554; https://doi.org/10.3390/info14100554 - 10 Oct 2023
Viewed by 1228
Abstract
Top-Down models are defined by hardware architects to provide information on the utilization of different hardware components. The target is to isolate the users from the complexity of the hardware architecture while giving them insight into how efficiently the code uses the resources. [...] Read more.
Top-Down models are defined by hardware architects to provide information on the utilization of different hardware components. The target is to isolate the users from the complexity of the hardware architecture while giving them insight into how efficiently the code uses the resources. In this paper, we explore the applicability of four Top-Down models defined for different hardware architectures powering state-of-the-art HPC clusters (Intel Skylake, Fujitsu A64FX, IBM Power9, and Huawei Kunpeng 920) and propose a model for AMD Zen 2. We study a parallel CFD code used for scientific production to compare these five Top-Down models. We evaluate the level of insight achieved, the clarity of the information, the ease of use, and the conclusions each allows us to reach. Our study indicates that the Top-Down model makes it very difficult for a performance analyst to spot inefficiencies in complex scientific codes without delving deep into micro-architecture details. Full article
(This article belongs to the Special Issue Advances in High Performance Computing and Scalable Software)
Show Figures

Figure 1

29 pages, 1605 KiB  
Article
Energy-Efficient Parallel Computing: Challenges to Scaling
by Alexey Lastovetsky and Ravi Reddy Manumachu
Information 2023, 14(4), 248; https://doi.org/10.3390/info14040248 - 20 Apr 2023
Cited by 3 | Viewed by 1860
Abstract
The energy consumption of Information and Communications Technology (ICT) presents a new grand technological challenge. The two main approaches to tackle the challenge include the development of energy-efficient hardware and software. The development of energy-efficient software employing application-level energy optimization techniques has become [...] Read more.
The energy consumption of Information and Communications Technology (ICT) presents a new grand technological challenge. The two main approaches to tackle the challenge include the development of energy-efficient hardware and software. The development of energy-efficient software employing application-level energy optimization techniques has become an important category owing to the paradigm shift in the composition of digital platforms from single-core processors to heterogeneous platforms integrating multicore CPUs and graphics processing units (GPUs). In this work, we present an overview of application-level bi-objective optimization methods for energy and performance that address two fundamental challenges, non-linearity and heterogeneity, inherent in modern high-performance computing (HPC) platforms. Applying the methods requires energy profiles of the application’s computational kernels executing on the different compute devices of the HPC platform. Therefore, we summarize the research innovations in the three mainstream component-level energy measurement methods and present their accuracy and performance tradeoffs. Finally, scaling the optimization methods for energy and performance is crucial to achieving energy efficiency objectives and meeting quality-of-service requirements in modern HPC platforms and cloud computing infrastructures. We introduce the building blocks needed to achieve this scaling and conclude with the challenges to scaling. Briefly, two significant challenges are described, namely fast optimization methods and accurate component-level energy runtime measurements, especially for components running on accelerators. Full article
(This article belongs to the Special Issue Advances in High Performance Computing and Scalable Software)
Show Figures

Figure 1

20 pages, 752 KiB  
Article
Realizing Mathematics of Arrays Operations as Custom Architecture Hardware-Software Co-Design Solutions
by Ian Andrew Grout and Lenore Mullin
Information 2022, 13(11), 528; https://doi.org/10.3390/info13110528 - 04 Nov 2022
Cited by 1 | Viewed by 1933
Abstract
In embedded electronic system applications being developed today, complex datasets are required to be obtained, processed, and communicated. These can be from various sources such as environmental sensors, still image cameras, and video cameras. Once obtained and stored in electronic memory, the data [...] Read more.
In embedded electronic system applications being developed today, complex datasets are required to be obtained, processed, and communicated. These can be from various sources such as environmental sensors, still image cameras, and video cameras. Once obtained and stored in electronic memory, the data is accessed and processed using suitable mathematical algorithms. How the data are stored, accessed, processed, and communicated will impact on the cost to process the data. Such algorithms are traditionally implemented in software programs that run on a suitable processor. However, different approaches can be considered to create the digital system architecture that would consist of the memory, processing, and communications operations. When considering the mathematics at the centre of the design making processes, this leads to system architectures that can be optimized for the required algorithm or algorithms to realize. Mathematics of Arrays (MoA) is a class of operations that supports n-dimensional array computations using array shapes and indexing of values held within the array. In this article, the concept of MoA is considered for realization in software and hardware using Field Programmable Gate Array (FPGA) and Application Specific Integrated Circuit (ASIC) technologies. The realization of MoA algorithms will be developed along with the design choices that would be required to map a MoA algorithm to hardware, software or hardware-software co-designs. Full article
(This article belongs to the Special Issue Advances in High Performance Computing and Scalable Software)
Show Figures

Figure 1

Planned Papers

The below list represents only planned manuscripts. Some of these manuscripts have not been received by the Editorial Office yet. Papers submitted to MDPI journals are subject to peer-review.

Title: When does large scale computing with data generate value
Authors: Kartik B. Ariyur
Affiliation: School of Mechanical Engineering Purdue University
Abstract: Large scale computing problems of engineering relevance, whether it is search, constraint satisfaction, estimation, or resource allocation, can all be mapped into problems of optimization. Large dimensional resource allocation problems are extremely sensitive to uncertainties in model parameters. The algorithms that estimate these parameters, including neural networks or GANs need the data to be sampled fast enough, stationarity of some properties such as chess rules, constraints on the underlying physical system, observability of physical state, identifiability of parameters, and have adequate signal to noise ratio for reliable performance. Various factors, including uncertainties in components and operating environment, noise, disturbances, corruption or disruption of measurements by competing stake holders prevent one or more of these conditions from being satisfied. This can make the estimation or resource allocation problems ill posed, or greatly increase the energy and time required for obtaining good data. Similarly, many large data sets collected by industry or governments may not satisfy these properties and assuming they do could generate catastrophic results.

Back to TopTop