Supercomputing Frontiers Europe 2019
NASA Ames Research Center, USA
Title:Advanced Computing at NASA
Dr. Rupak Biswas is currently the Director of Exploration Technology at NASA Ames Research Center, Moffett Field, Calif., and has held this Senior Executive Service (SES) position since January 2016. In this role, he in charge of planning, directing, and coordinating the technology development and operational activities of the organization that comprises of advanced supercomputing, human systems integration, intelligent systems, and entry systems technology. The directorate consists of approximately 700 employees with an annual budget of $160 million, and includes two of NASA’s critical and consolidated infrastructures: arc jet testing facility and supercomputing facility. He is also the Manager of the NASA-wide High End Computing Capability Project that provides a full range of advanced computational resources and services to numerous programs across the agency. In addition, he leads the emerging quantum computing effort for NASA. Dr. Biswas received his Ph.D. in Computer Science from Rensselaer in 1991, and has been at NASA ever since. During this time, he has received several NASA awards, including the Exceptional Achievement Medal and the Outstanding Leadership Medal (twice). He is an internationally recognized expert in high performance computing and has published more than 160 technical papers, received many Best Paper awards, edited several journal special issues, and given numerous lectures around the world.
University of California Berkeley, USA
Tentative Title: Memristor – Remembrance of things Past
Leon Chua is widely known for his invention of the Memristor and the Chua’s Circuit. His research has been recognized internationally through numerous major awards, including 17 honorary doctorates from major universities in Europe and Japan, and 7 USA patents. He was elected as Fellow of IEEE in 1974, a foreign member of the European Academy of Sciences (Academia Europea) in 1997, a foreign member of the Hungarian Academy of Sciences in 2007, and an honorary fellow of the Institute of Advanced Study at the Technical University of Munich, Germany in 2012. He was honored with many major prizes, including the Frederick Emmons Award in 1974, the IEEE Neural Networks Pioneer Award in 2000, the first IEEE Gustav Kirchhoff Award in 2005, the International Francqui Chair (Belgium) in 2006, the Guggenheim Fellow award in 2010, Leverhulme Professor Award (United Kingdom) during 2010-2011, and the EU Marie curie Fellow award, 2013.
Prof. Chua is widely cited for the 12 hugely popular lectures he presented at the hp Chua Lecture Series, entitled “From Memristors and Cellular Nonlinear Networks to the Edge of Chaos”, during the fall of 2015, and now accessible through YouTube via the link.
Argonne National Laboratory, USA
Tentative Topic: Exascale Supercomputing
Paul Messina is Director of Argonne National Laboratory’s Computational Science Division and an Argonne Distinguished Fellow. During 2015-2017, he served as Project Director for the U.S. Department of Energy Exascale Computing Project. From 2008-2015, he was Director of Science for the Argonne Leadership Computing Facility.
From 1987-2002, he was founding Director of the Center for Advanced Computing Research at the California Institute of Technology (Caltech), PI for the CASA gigabit network testbed, Chief Architect for NPACI, and co-PI for the National Virtual Observatory and TeraGrid. In 1990, he conceived and led the Concurrent Supercomputing Consortium that created and operated the Intel Touchstone Delta, at the time the world’s most powerful scientific computer and in 1999-2000 led the DOE-NNSA ASCI program while on leave from Caltech.
During 1973-1987, he held a number of positions in the Applied Mathematics Division and was founding Director of the Mathematics and Computer Science Division at Argonne.
University of Tsukuba, Japan
Title: What’s the next step of accelerated supercomputing?
Taisuke Boku received Master and PhD degrees from Department of Electrical Engineering at Keio University. After his career as assistant professor in Department of Physics at Keio University, he joined to Center for Computational Sciences (former Center for Computational Physics) at University of Tsukuba where he is currently the deputy director, the HPC division leader and the system manager of supercomputing resources. He has been working there more than 20 years for HPC system architecture, system software, and performance evaluation on various scientific applications. In these years, he has been playing the central role of system development on CP-PACS (ranked as number one in TOP500 in 1996), FIRST (hybrid cluster with gravity accelerator), PACS-CS (bandwidth-aware cluster) and HA-PACS (high-density GPU cluster) as the representative supercomputers in Japan. Also, He is a member of system architecture working group of Post-K Computer development. He is a coauthor of ACM Gordon Bell Prize in 2011.
In this talk, I will present our current research on advanced hybrid cluster computing with the latest FPGA technology to provide flexible and power-effective accelerated computing in next generation. We apply FPGA technology not just for computation but also for interconnection for large scale parallel systems without communication bottleneck. The concept is based on the high performance optical interconnection interface suited in the latest version of FPGA with up to 100Gbps speed on multiple communication ports. We have been developing a prototype system of such a concept named AiS (Accelerator in Switch) to gather the computation and communication on FPGA device with flexible programming based on OpenCL instead of traditional Verilog HDL, for user friendly FPGA programming.
Beside of FPGA use in clusters, we still depend on GPU with its excellent performance on problems or partial problems Where it works well. In some applications, there is parts where there is not enough parallelism to use GPU efficiently or irregular computation not fit to GPU’s SIMT computation manner to cause the performance bottleneck especially on strong scaling. Therefore, we introduce GPU and FPGA together as different type of accelerators.
In this talk, the brief image and prototype design/implementation of AiS-based system developed in Center for Computational Sciences, University of Tsukuba is presented. And our new machine based on this concept which will be delivered on March 2019, named Cygnus, is introduced and several preliminary performance evaluation on actual applications is presented.
University of Leeds, UK
Title: Why Topology is Necessary at Exascale – And Why it’s Not Easy
Dr. Hamish Carr is a Senior Lecturer in the School of Computing at the University of Leeds, and is one of the principal researchers in the field of computational topology and the mathematical foundations of scientific visualization. His thesis work included the standard efficient serial algorithm and visualizations based on the topological structure called the contour tree, and he has been involved in developing, refining and applying these tools to a variety of data analytic problems ever since. He has also worked on accelerations to direct volume rendering, improvements in computational statistics for visualization, and tools for direct visualization of two or more functions rather than the single function normally assumed. In recent years, he has been collaborating with Los Alamos and Lawrence Berkeley National Laboratories on developing topological tools for shared-memory and distributed parallel machines, and has contributed to the new vtk-m multicore visualization toolkit under the DoE ALPINE project.
As computation scales to exabyte levels, humans are increasingly unable to process the data directly. In a realistic assessment, a human might meaningfully perceive 10 GB of data in a lifetime, which requires billion-fold data reduction without information loss. As a result, techniques like computational topology, with mathematically well-founded approaches, become increasingly necessary to analyse and display (i.e. visualise) extreme data sets. However, much of the mathematics on which these methods are based is either linearly defined or couched in serial metaphors. As a result, one of the major challenges to exascale comprehension is now the ability to scale computational topology to massive hybrid systems using both shared-memory and distributed parallelism. This talk will therefore give an overview of the basic ideas behind computational topology at scale, and illustrate recent approaches to scalability.
Argonne National Laboratory, USA
Title: Understanding Cities through Measurement and Embedded Intelligence
Charlie Catlett is the founding director of the Urban Center for Computation and Data, UrbanCCD, which brings social, physical, and computational
scientists together with artists, architects, technologists, and policy makers to explore urban dynamics and design. He is also a Senior Computer Scientist at Argonne National Laboratory and a Senior Fellow at the Mansueto Institute for Urban Innovation at The University of Chicago.
From 2007 to 2011 he was the Chief Information Officer at Argonne National Laboratory, and from 2004 to 2007 he was Director of the National Science Foundation’s TeraGrid initiative – a nationally distributed supercomputing facility involving fifteen universities and federal laboratories. From 1999 to 2004 Charlie directed the design and deployment of I-WIRE, a dedicated fiber optic network funded by the State of Illinois, which connects research institutions in the Chicago area and downstate Illinois to support advanced research and education. In recent years he has created a new form of urban measurement infrastructure, the “Array of Things,” implementing new sensing and edge computing technologies.
Before joining the University of Chicago and Argonne in 2000, Charlie was Chief Technology Officer at the National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign. Beginning at NCSA’s founding in 1985 he participated in the development of NSFNET, one of several early national networks that evolved into what we now experience as the Internet. During the exponential growth of the web following the release of NCSA’s Mosaic web browser, his team developed and supported NCSA’s scalable web server infrastructure.
Recognized one of 25 “Doers, Dreamers & Drivers” of 2016 by Government Technology magazine and in 2014 as one of Chicago’s “Tech 50” technology leaders by Crain’s Chicago Business, Charlie is a Computer Engineering graduate of the University of Illinois at Urbana-Champaign.
Abstract: Urbanization is one of the great challenges and opportunities of this century, inextricably tied to global challenges ranging from climate change to sustainable use of energy and natural resources, and from personal health and safety to accelerating innovation and education. There is a growing science community—spanning nearly every discipline—pursuing research related to these challenges. For many urban questions, however, there is a need for measurements with greater spatial and/or temporal resolution than is currently available for understanding air quality, microclimate, vibration, noise, and other factors. Concurrently, many measurements—from the flow of people through a public space to the impact of at-grade rail crossings on emergency response—require more sophisticated “measurements” involving embedded, or “edge” computing within the urban infrastructure. Catlett will discuss the Array of Things initiative that Argonne National Laboratory and the University of Chicago are doing in partnership with the City of Chicago and a growing number of cities around the world, involving novel remotely programmable embedded measurement systems. Catlett will also discuss efforts to build computational models encompassing multiple facets of cities, ranging from transportation to weather to energy to social-economic factors.
IBM Research Labs Zurich, Switzerland
Title: “In-memory Computing”: Accelerating AI Applications
Evangelos Eleftheriou, received his Ph.D. degree in Electrical Engineering from Carleton University, Ottawa, Canada, in 1985. In 1986, he joined the IBM Research – Zurich laboratory in Rüschlikon, Switzerland, as a Research Staff Member. Since 1998, he has held various management positions and is currently responsible for the neuromorphic computing activities of IBM Research – Zurich. His research interests include signal processing, coding, non-volatile memory technologies and emerging computing paradigms such as neuromorphic and in-memory computing for AI applications. He has authored or coauthored over 200 publications, and holds over 160 patents (granted and pending applications). In 2002, he became a Fellow of the IEEE. He was co-recipient of the 2003 IEEE Communications Society Leonard G. Abraham Prize Paper Award. He was also co-recipient of the 2005 Technology Award of the Eduard Rhein Foundation. In 2005, he was appointed an IBM Fellow. The same year he was also inducted into the IBM Academy of Technology. In 2009, he was co-recipient of the IEEE CSS Control Systems Technology Award and of the IEEE Transactions on Control Systems Technology Outstanding Paper Award. In 2016, he received an honoris causa professorship from the University of Patras, Greece. In 2018, he was inducted into the US National Academy of Engineering as Foreign Member.
In today’s computing systems based on the conventional von Neumann architecture, there are distinct memory and processing units. Performing computations results in a significant amount of data being moved back and forth between the physically separated memory and processing units. This costs time and energy, and constitutes an inherent performance bottleneck. It is becoming increasingly clear that for application areas such as AI (and indeed cognitive computing in general), we need to transition to computing architectures in which memory and logic coexist in some form. Brain-inspired neuromorphic computing and the fascinating new area of in-memory computing or computational memory are two key non-von Neumann approaches being researched. A critical requirement in these novel computing paradigms is a very-high-density, low-power, variable-state, programmable and non-volatile nanoscale memory device. There are many examples of such nanoscale memory devices in which the information is stored either as charge or as resistance. However, one particular example is phase-change-memory (PCM) devices, which are very well suited to address this need, owing to their multi-level storage capability and potential scalability.
In in-memory computing, the physics of the nanoscale memory devices, as well as the organization of such devices in cross-bar arrays, are exploited to perform certain computational tasks within the memory unit. I will present how computational memories accelerate AI applications and will show small- and large-scale experimental demonstrations that perform high-level computational primitives, such as ultra-low-power inference engines, optimization solvers including compressed sensing and sparse coding, linear solvers and temporal correlation detection. Moreover, I will discuss the efficacy of this approach to efficiently address not only inferencing but also training of deep neural networks. The results show that this co-existence of computation and storage at the nanometer scale could be the enabler for new, ultra-dense, low-power, and massively parallel computing systems. Thus, by augmenting conventional computing systems, in-memory computing could help achieve orders of magnitude improvement in performance and efficiency.
Anne C. Elster
Norwegian University of Science and Technology, Norway
Title: Supercompting and AI: Impact and Opportunities
Dr. Anne C. Elster (PhD Cornell Univ. 1984) , is a Professor at the Department of Computer Science (IDI) at Norwegian Univ. of Science & Technology (NTNU) in Trondheim Norway, where she established the IDI/NTNU HPC-Lab, a well-respected research lab in heterogeneous computing. She also holds a current Visiting Scientist position at the Univ. of Texas at Austin, USA.
Her current research interests are in high-performance parallel computing, currently focusing on developing good models and tools for heterogeneous computing and parallel software environments including methods that applying machine learning for code optimization and image processing, as well as developing parallel scientific codes that interact visually with the users by taking advantage of the powers in modern GPUs. Her novel fast linear bit-reversal algorithm from 1987 is still noteworthy.
She has given several invited talks and has been active participant and committee member of ACM/IEEE SC (Supercomputing) since 1990. She also served on MPI 1&2, is a Senior member of IEEE, Life member of AGU (American Geophysical Unions, as welland as a member of ACM, SIAM and Tekna. Funding partners/collaborators include EU H2020, The Research Council of Norway, AMD, ARM, NVIDIA, Statoil and Schlumberger.
Supercomputing is still needed in order to be able to get the desired performance for computational intensive tasks ranging from Big Data to astro-physics and weather simulations. Traditionally the HPC field drove companies like Cray, IBM and others to develop processors for supercomputing. However, the market forces in other fields have since the proliferation of COTS (commercial off-the shelf) processors, including GPUs for gaming and now more recently AI, driven the innovation in processor design, This means that algorithms, tools and applications now should adapt and take advantage of tensor processors, Machine Learning techniques, and other related technology, rather than expect that old computational models with hold true. As will be discussed in this talk, this also an opportunity to help develop better graph algorithms for AI as well as apply some of the techniques from AI to HPC challenges.
ETH Zürich, Switzerland
Tentative Topic: Extreme Scale Graphs
Torsten is an Associate Professor of Computer Science at ETH Zürich, Switzerland. Before joining ETH, he led the performance modeling and simulation efforts of parallel petascale applications for the NSF-funded Blue Waters project at NCSA/UIUC. He is also a key member of the Message Passing Interface (MPI) Forum where he chairs the “Collective Operations and Topologies” working group. Torsten won best paper awards at the ACM/IEEE Supercomputing Conference SC10, SC13, SC14, EuroMPI’13, HPDC’15, HPDC’16, IPDPS’15, and other conferences. He published numerous peer-reviewed scientific conference and journal articles and authored chapters of the MPI-2.2 and MPI-3.0 standards. He received the Latsis prize of ETH Zurich as well as an ERC starting grant in 2015. His research interests revolve around the central topic of “Performance-centric System Design” and include scalable networks, parallel programming techniques, and performance modeling. Additional information about Torsten can be found on his homepage at htor.inf.ethz.ch.
National Center for Supercomputing Applications, USA
Title: Driving innovation at the interface of high performance computing and artificial intelligence for Multi-Messenger Astrophysics.
Founder and Lead of the Gravity Group at the National Center for Supercomputing
Applications, University of Illinois at Urbana-Champaign. Eliu obtained a Master in Applied
Mathematics and Theoretical Physics, followed by a PhD in Theoretical Astrophysics, at the
University of Cambridge, United Kingdom.
His work is motivated by grand computational challenges in the nascent field of
Multi-Messenger Astrophysics. He combines his expertise in theoretical astrophysics and
large scale computing to create scenarios for Multi-Messenger Astrophysics, driving
innovation on the exploitation of machine and deep learning for the modeling, detection and
scientific exploitation of transient astrophysical phenomena.
Infinite Curation, Japan
Title: Sequence Similarity Search for Large-scale Metagenomic Data using Liquid Immersion Cooling Supercomputer
Yoshinori Kimura is a Senior Software Engineer of Infinite Curation (Seconded from group company, ExaScaler). He has 15 years experience of image processing, volume visualization and customer oriented application in medical field.
Mr. Kimura started his career at TeraRecon as co-creator of AquariusNET, the world’s first commercial streaming, thin-client, enterprise 3D imaging server. After the design and development of basic framework of Aquarius iNtuition, which is next generation of AquariusNET, he concentrated to develop more academic function such like MRI ejection fraction, flow, tractography analysis. The radiologists and clinicians highly evaluated the ease of use and stability of both AquariusNET and Aquarius iNtuition, thousands of systems are installed in the world.
Previously Mr. Kimura developed interface between cloud service and quantum computer simulator on PEZY-SCnp/ZettaScaler1.6 system at ExaScaler since he has experience of parallel processing systems in his postgraduate course. And he seconded to Infinite Curation which is group company of ExaScaler to develop medical/genomic application onto PEZY-SC/ZettaScaler system. His design and analysis skills obtained in his experience, flexibility to learn various fields are suitable for this purpose. He received the Bachelor’s degree in Mechanical Engineering, and the Master’s degree in Information Science from Tohoku University in 1998 and 2000.
University of California San Diego, USA
Tentative Title: Computational challenges in understanding the structure and function of the microbiome
Tomasz Kościółek is a postdoctoral research associate in Rob Knight’s Lab at the University of California San Diego. He holds an MSc in chemistry (2010) from the Jagiellonian University in Kraków, Poland; and a PhD in biological sciences (2015) from University College London, United Kingdom. His doctoral work was focused on the application of sequence covariation methods for protein structure prediction and computational predictions of dynamic properties in intrinsically disordered proteins. Since he joined the Knight Lab, he focuses on large-scale predictions of the structure and function of microbial proteins (lead researcher on Microbiome Immunity Project – collaboration involving IBM World Community Grid), the development of scalable computational methods for microbiome analyses (e.g. QIIME 2 and Qiita) and studies of the role of the microbiome in health and disease, especially in the context of mental health.
Ayar Labs, USA
Title: Optical communications for high-bandwidth chip I/O in computing systems
Alex Wright-Gladstein is co-founder of Ayar Labs, a company that is making computing faster and more energy efficient by using light to move data between chips instead of electricity. She founded the company while at MIT, where she met the inventors of the first processor to communicate using light. Prior to Ayar Labs, Alex was the Energy Entrepreneurship Practice Leader at MIT, an Energy Markets Specialist and Program Manager at EnerNOC, the first smart grid company, and the founder of Everyday Science.
École polytechnique fédérale de Lausanne, Switzerland
Title: Studying brain structure and function through detailed computational modelling and topological analysis
Michael Reimann is a senior staff scientist in the Blue Brain Project at the EPFL, Switzerland. For his doctoral research he developed algorithms to derive microcircuit connectivity and study the emergence of microcircuit activity.
He is one of the main contributors of digital microcircuit modeling and simulation efforts in the Blue Brain Project. His research is focused on synaptic connectivity at all scales, how it is shaped by plasticity mechanisms and how this in turn determines brain function. To this end he employs advanced simulation tools on massively parallel supercomputing systems and develops novel analyses, based on classical information theory and algebraic topology.
The function of the brain is thought to be determined by the structure of its synaptic connections. Unfortunately, the study of the nature of this relation has been hampered by an incomplete knowledge about the synaptic wiring of neurons on the local scale of a microcircuit or brain region. Through the use of massively parallel supercomputing architectures, we were able to build, simulate and analyze the most detailed model of a rat brain region to date, that is based on the available biological data and principles, and extensively validated. Using methods from algebraic topology, we characterized highly non-random structures in the resulting predicted synaptic connectivity in the form of all-to-all connected, directed motifs. When we simulated the electrical activity of the modeled neurons we found that the topologically identified motifs led to correlated neuronal activity, thus furthering our understanding of the link between synaptic structure and neuronal function. Running the model on the MIRA system at Argonne National Labs also allowed us to study how neuronal activity in turn re-shapes connectivity through mechanisms of synaptic plasticity, thus completing a loop from structure to function and back again.