News related to recent market trend,Investor,research,travelling,engineering and books.

ads

Olymp the fast calculator: Hardware becomes more energy-efficient and increasingly supports methods of artificial i...

HIGH PERFORMANCE COMPUTING

Olymp the fast calculator:

Hardware becomes more energy-efficient and increasingly supports methods of artificial intelligence. This makes it possible to simulate increasingly complex and, above all, faster simulations of complex processes.

Hazel Hen: The system at the highest performance center in Stuttgart is 95% loaded by customer orders. (Photo: Boris Lehrner for HLRS )
The Top500 list of the most powerful supercomputers is still the benchmark for success in high-performance computing (HPC). The list, created twice a year, is based on a performance benchmark called Linpack. In the meantime, however, there are also other benchmarks, such as the HPCG benchmark, which is based on typical storage operations of commercial software, or Green500, where the computing power per watt is consumed.

The fastest technological advances in the field of computing accelerators play an important role in the top group in the latter segment: graphics processors (GPUs), which facilitate the parallel processing of data and at the same time have lower power consumption than conventional processors. The manufacturer Nvidia is the driving force behind the GPU market. The Tesla-P100-GPU and processors from Intel have given the Swiss supercomputer "Piz Daint" the necessary power to reach the top three (see box).

However, this acceleration must also be reflected in the field of lines between processors, GPUs, and memory. For two years now, Intel's Omni-Path Architecture (OPA) has enjoyed great popularity. It offers high data throughput of 100 Gbit / s on the PCIe bus, low latency and low power consumption in compatible devices. This success points to a growing problem of high-performance technology: pure computing is now usually fast enough, but the connection between the individual components, which is still based on the standards Ethernet and Infiniband, is lagging behind. The PCIe bus 3.0 has not been renewed for many years, only since June 2017 there are new specifications.They should allow up to 128 Gbit / s in full-duplex mode.
The search for energy-saving computer architectures in the various national and university HPC projects has one reason: the enormous energy consumption that future systems will have. Exceptional computing power on the exaflops scale means trillion computing operations per second and high cooling requirements. The industry calls this goal an exascale. "Such systems have an energy requirement of 20 MW to 30 MW", explains Klaus Gottschalk, HPC expert at IBM Germany. This need is already covered today, but at an unacceptable cost. Gottschalk sees therefore an Exascale architecture only from the year 2020 or even only 2023 come.
Michael Resch, Head of the Stuttgart High-Performance Calculation Center (HLRS) and Chairman of the Gauss Center for Supercomputing, does not see any meaning in the Exascale race. Its Cray-XC40 supercomputer "Hazel Hen" delivers 7,42 Petaflops top performers, but the industry and business customers who use the computer wants a real performance, which is currently about 1 Petaflops. Reliability and availability are as important as pure computing performance, according to Resch. In addition, numerous raking projects are carried out simultaneously and in parallel. "Chinese supercomputers are only 15% to 20% loaded according to Chinese data," Resch said at the International Supercomputing Conference in Frankfurt / M. find out."Our Hazel Hen is about 95% loaded all the time."

HPC has a wealth of application fields. At the HLRS, this is primarily the material science. But also the meteorologists profit. For example, in order to better prosecute storms or predict the occurrence of storms, experts in the USA use methods of artificial intelligence (AI). A new model for global weather forecasting is being developed by the University Corporation for Atmospheric Research (UCAR), the National Center for Atmospheric Research (NCAR) and the IBM subsidiary Weather.com. Data volumes of up to 1 petabyte are processed in the main memory using appropriate algorithms and models. The goal is to make predictions more accurate in smaller and smaller regions. The simulation of global climate change can also be realized with these methods,

Nvidia has developed its new V100-GPU especially for applications of the AI. First specimens were handed over to 15 AI research facilities last week. These include the German Research Center for Artificial Intelligence (DFKI) and the Max Planck Institute for Intelligent Systems.The GPU has 21 billion transistors. At Princeton University, nuclear fusion researchers want to use these chips to achieve their accuracy for predicting plasma disturbances of up to 95%. Also the demand of the industry for AI is high, Eg for calculations and simulations in the search for raw materials such as oil and gas.
While the hardware can make significant progress, the software must also keep up. Not every application is suitable for every HPC environment, since the prerequisite is always the parallel processing of data in a computing cluster. HPC has been growing since 2015 with the analysis of Big Data to HPDA (High Performance Data Analytics). At HPDA, the use of AI algorithms is now essential. With regard to this field, Cray presented an Analytics software for Cray-XC supercomputers at the ISC conference with Urika-XC.
The suite includes tools for graph analysis, AI technology, and big-data analysis. "This allows analysis and AI workloads to be run parallel to scientific modeling and simulations, eliminating the time and cost for data migration between systems," says the manufacturer. The solution is already used on the Swiss system Piz Daint, but also on the HLRS.

0 comment: