Plenary Session

Conference Opening
Vladimir Voevodin, Moscow State University, Russia

An Overview of High Performance Computing and Future Requirements
Jack Dongarra, University of Tennessee, Oak Ridge National Laboratory, and University of Manchester, USA

In this talk we examine how high performance computing has changed over the last 10-year and look toward the future in terms of trends. These changes have had and will continue to have a major impact on our numerical scientific software. A new generation of software libraries and algorithms are needed for the effective and reliable use of (wide area) dynamic, distributed and parallel environments.  Some of the software and algorithm challenges have already been encountered, such as management of communication and memory hierarchies through a combination of compile--time and run--time techniques, but the increased scale of computation, depth of memory hierarchies, range of latencies, and increased run--time environment variability will make these problems much harder. 

HLRS - A National Supercomputing Center for Research and Industry
Michael Resch, University of Stuttgart, Germany

Established in 1996 HLRS was the first German national supercomputing center. Over more than 20 years its focus and activities have been shaped by a changing landscape of research and industry. In this talk we will present a number of questions relevant for any HPC center. What are the driving technological forces for HPC? What kind of research questions arise in HPC? How does industry change the operation of HLRS? What are the upcoming new challenges like Machine Learning and Artificial Intelligence? What impact will they have on HPC?

An Overview of Post-K Supercomputer
Yutaka Ishikawa, RIKEN, Japan

The post-K, a flagship supercomputer in Japan, is being developed by Riken and Fujitsu. It will be the first supercomputer with an Armv8-A with SVE (Scalable Vector Extension) architecture. It will consist of more than 150k nodes connected by the TofuD network, an enhanced version of Tofu interconect used in K computer. The general operation will start in 2021. In this presentation, an overview of the post-K hardware and its software stack will be presented.

Best Practices for Implementing Supercomputer Modeling in the Development of Materials
Lief Pedersen, BIOVIA, Dassault Systemes, USA

Developing any modern material—from product packaging to arospace composites—requires careful study of the physical processes occurring inside the material. Unfortunately, in many cases, research data methods significantly limit laboratory studies related to materials development, because even the most powerful equipment does not allow materials scientists to observe processes at the atomic level.

Since the 1970s, computer modeling has provided an auxiliary method for studying processes at the atomic level to reduce the time it takes to develop materials. Today we live in the era of digital supercomputer modeling, which allows scientists to investigate processes throughout the entire lifecycle of a material—from initial concept to bringing a finished product to market—with reduced time, cost and effort.

Leif Pedersen will present the experiences, effectiveness and results obtained by world-famous companies utilizing computer-aided modeling tools to characterize material properties, with an emphasis on the use of high-performance computers in conducting such research.

High performance computing and machine learning in hydrocarbon exploration and recovery problems
Sergey Tikhotskiy, RAS, Russia

Geophysics has traditionally been one of the main consumers of computing resources throughout the world. In recent decades, the tasks of processing geophysical information have been supplemented by the problems of complex field modeling: the processes of generation and migration of hydrocarbons, multiphase filtration, changes in the stress-strain state, and oil recovery intensification. These problems are of particular relevance because of the need to engage in the recovery of untraditional hydrocarbon reserves. All the above tasks cannot be solved without the use of high-performance computing using supercomputers and modern data analysis methods.

In hydrocarbon exploration, it is necessary to develope the high-performance algorythms and software to refine the structure and evaluate the reservoir rocks properties. In particular, this includes methods for processing modern wide-azimuth and multicomponent seismic data, including migration and seismic inversion for anisotropic media, as well as full-waveform inversion. These  are based on the multiple calculation of the seismic wavefield in three-dimensional inhomogeneous anisotropic media, as well as nonlinear multiparameter optimization. To carry out such calculations in a typical field model, with only one position of the seismic source, it is necessary to have 3 petabytes of RAM and perform 1021 operations with a floating point (i.e. 1000 exaflops). It is also valuable to be able to estimate the effective physical properties of micro-inhomogeneous porous-fractured media at different scales, which is necessary for a correct estimation of the transport prorerties from logs and field geophysical data. Correct assessment of fluid saturation also requires the development of methods for modeling and inversion of the electromagnetic field for such media. The maximum task is to develop algorithms for the combined inversion of various physical fields with detail that is adequate to the needs of the exploration industry. Because the evaluation of the collector and transport properties and other parameters of the geological environment according to the data of geophysical studies is a non-linear and incorrect inverse problem, for its solution it is effective to apply the methods of big data analysis and machine learning.

The imortant task in the recovery process modeling is the transition from the traditional approach, in which hydrodynamic and geomechanical modeling of reservoirs are performed independently, to their conjugate modeling, since a change in the stress-strain state leads to a change in the transport properties of rocks. This problem leads to the complex numerical schemes that require the use of special grids and are unstable. Overcoming the arising difficulties and creating practically applicable software products is also possible solely on the basis of high-performance computing.

A special class of problems that require the use of high-performance computing is the design and optimization of methods for oil recovery enchancement: hydraulic fracturing, steam injection, method of in-situ combusion, etc.

The main goal of the simulation is the digital model of the field, including information on the current state of the reservoir, its stress-strain state, fluid flow. Such a model should be updated and refined in real time, taking into account the newly acquired exploration, well and monitoring data. Real-time analysis of the simulation results, using artificial intelligence methods, must lead to the optimisation of the recovery scheme, drilling parameters and use of the enhanced oil recovery methods. This functionality leas to the concept of "smart field". It is advisable to develop a typical supercomputer platform (including hardware and software), which could then be scaled and applied in all hydrocarbon fields as a standard means for the operation of the described digital model.

Quantum information technologies: current status and prospects of their applications
Vladimir Gerdt, JINR, Russia

The talk based on materials of the workshop «Quantum Computing for High Energy Physics» (CERN, November 5-6, 2018) and other open sources. and contains the state-of-the-art brief review of quantum computing and quantum information. In consideration of the especially promising applications of quantum computing we place emphasis on the quadratic unconstrained binary optimization problems (quadratic programming) which form the basis of rapidly developing area - quantum machine learning  - perspective in analysis of big data.

R&D of a Quantum-Annealing Assisted Next-Generatioon HPC Infrastructure for computer science, data science and their fusion applications development
Hiroaki Kobayashi, Tohoku University, Japan

As the silicon technology driven by Moore’s law is facing the physical limitation, we are now moving to the post-Moore’s era in the design of high-performance computing architectures.   Quantum Annealing is one of  emerging information processing technologies in the Post-Moore’s era and is  expected to work well for combinatorial optimization problems.   In my talk, I will be presenting our on-going project entitled Quantum-Annealing assisted next generation HPC infrastructure that aims to integrated quantum annealing information processing into a conventional HPC system as  an accelerator of  combinatorial optimization problems.  I am also discussing the design of several applications that integrate computational science and data science approaches.

Evgeny Tyrtyshnikov, MSU, RAS, Russia

Presentation by T-Platforms

Presentation by Mellanox

Presentation by IBM

Presentation by Dell

Presentation by Xilinx

Presentation by AMD