Feed aggregator

Nyriad to Demonstrate New Operating System for the SKA Precursor MWA Telescope at SC17

HPC Wire - Wed, 11/15/2017 - 22:35

DENVER, Nov. 14, 2017— Exascale computing company Nyriad Limited, the first commercial spin-out from the SKA, has been partnering with the International Centre of Radio Astronomy Research (ICRAR) to design, develop and deploy a ‘Science Data Processing’ (SDP) operating system for the SKA-Low precursor telescope, the Murchison Widefield Array (MWA).

Nyriad will be demonstrating the first commercial product from the collaboration called Nsulate, a GPU-accelerated storage solution which enables increased storage resilience while reducing the storage power requirements by over 50 percent.

ICRAR are engaged in several aspects of the preconstruction phase for the SKA program, including research, software engineering and data intensive astronomy, and are collaborating with Nyriad on developing a GPU-accelerated OS architecture for next-generation supercomputers. This new OS will eliminate the need for dedicated storage infrastructure while dramatically reducing power consumption and infrastructure costs. It will increase performance by using the GPUs traditionally dedicated to data processing alone to also perform the storage-processing functions for the supercomputer, thereby keeping the storage very close to the computing nodes. Nyriad is further collaborating with ICRAR on GPU-accelerating their Daliuge graph processing framework to analyse vast streams of radio antennae data in real-time.

Director of ICRAR’s Data Intensive Astronomy Program, Professor Andreas Wicenec, stated, “Nyriad was founded following discussions and consulting work around the SKA data challenges. The Nyriad founders, Matthew A. Simmons (CEO) and Alex St. John, identified a need for innovative approaches merging storage and processing which could benefit the SKA, but it became obvious that many of the ‘Big Data’ projects arising in other sciences, industries and governments would benefit as well.”

“The Universe is a big place,” said Nyriad CTO Alex St. John, “so we’ve been forced to rethink the entire software and hardware stack to come up with new computer designs that can handle the data processing volumes necessary to map the cosmos, Nsulate is the first of a suite of solutions that address these problems.” St. John is best known for his early pioneering work at

Microsoft on creating the Direct3D API and DirectX media OS that gave rise to modern consumer and HPC GPUs.

The new OS is being demonstrated at Nyriad partner booths TYAN and by HPC Systems at the SuperMicro booth at SC17, and at the Nyriad suite by appointment.

About Nyriad

Nyriad is a New Zealand-based exascale computing company specialising in advanced data storage solutions for big data and high performance computing. Born out of its consulting work on the Square Kilometre Array Project, the company was forced to rethink the relationship between storage, processing and bandwidth to achieve a breakthrough in system stability and performance capable of processing and storing over 160Tb/s of radio antennae data in real-time, within a power budget impossible with any modern IT solutions.


The International Centre for Radio Astronomy Research (ICRAR) was founded in August 2009 with the specific purpose of supporting Australia’s bid to host the world’s largest radio telescope and one of the largest scientific endeavors in history, the Square Kilometre Array (SKA). ICRAR is a joint venture between Curtin University and The University of Western Australia (UWA), with funding support from the State Government of Western Australia. ICRAR has research nodes at both universities and is now host to over 150 staff and postgraduate students.

Source: Nyriad, ICRAR

The post Nyriad to Demonstrate New Operating System for the SKA Precursor MWA Telescope at SC17 appeared first on HPCwire.

NEC Supplies Top500 Supercomputer to Johannes Gutenberg University Mainz in Germany

HPC Wire - Wed, 11/15/2017 - 19:21

DUSSELDORF and TOKYO, Nov. 15 – NEC Corporation today (Nov. 16 in Tokyo) announced that NEC Deutschland GmbH has delivered an LX  series supercomputer to Johannes Gutenberg University Mainz (JGU), one of Germany’s leading research universities and part of the German Gauss Alliance consortium of excellence in high-performance computing (HPC). The new HPC cluster ranks 65th in the most current TOP500 list of the fastest supercomputers in the world from November 2017 and 51st in the Green500 list of the most energy-efficient supercomputers.

This cluster extends the existing MOGON-II cluster, thereby providing a total computational capacity of approximately 1.9 Petaflop/sec. It offers high performance computing services for researchers at JGU and the Helmholtz Institute Mainz (HIM), a research institute specializing in high-energy physics and antimatter research. JGU is a member of the “Alliance for High-Performance Computing Rhineland-Palatinate” (AHRP) and offers access to MOGON-II to all universities in Rhineland-Palatinate.

The new MOGON-II HPC cluster upgrade consists of 1040 dual-socket compute nodes, each equipped with two Intel(R) Gold 6130 CPUs and a total memory of 122 TB.

The nodes are connected through a high-speed Intel Omni-Path network with a topology that allows continuous expansion of the system, which meets the ongoing growth of HPC demand from researchers from JGU and HIM.

The MOGON-II cluster is connected to a 5 PetaByte NEC LxFS-z parallel file-system capable of 80 GigaByte/s bandwidth. This highly innovative ZFS-based Lustre solution provides advanced data integrity features paired with a high density and high reliability design.

“We have been working together with NEC for many years now, and we are happy to confirm that this collaboration has always been very fruitful to our research members and to the excellence in research at Mainz University. The high sustained performance and stability of NEC’s HPC solution, as well as the dedication and skill of their team continuously deliver exceptional results,” emphasizes Professor Andre Brinkmann, Head of the Zentrum fur Datenverarbeitung and of the Efficient Computing and Storage Group at JGU.

“We are honored to see Johannes Gutenberg University Mainz and Helmholtz Institute Mainz, two highly respected members of the research community, adopt NEC’s latest HPC solution as part of extending the capabilities of the MOGON-II cluster,” said Yuichi Kojima, Vice President HPC EMEA at NEC Deutschland.

About Johannes Gutenberg University Mainz

With around 32,000 students and more than 4,400 academics and researchers from over 120 nations, Johannes Gutenberg University Mainz (JGU) is one of the largest research universities in Germany. Its main core research areas are in the fields of particle and hadron physics, the materials sciences, and translational medicine. The university campus is also home to four partner institutes involved in top-level non-university research: the Helmholtz Institute Mainz (HIM), the Max Planck Institute for Chemistry (MPI-C), the Max Planck Institute for Polymer Research (MPI-P), and the Institute of Molecular Biology (IMB).

About NEC Corporation 

NEC Corporation is a leader in the integration of IT and network technologies that benefit businesses and people around the world. By providing a combination of products and solutions that cross utilize the company’s experience and global resources, NEC’s advanced technologies meet the complex and ever-changing needs of its customers. NEC brings more than 100 years of expertise in technological innovation to empower people, businesses and society. For more information, visit NEC at http://www.nec.com.

Source: NEC Corp.

The post NEC Supplies Top500 Supercomputer to Johannes Gutenberg University Mainz in Germany appeared first on HPCwire.

Hyperion Market Update: ‘Decent’ Growth Led by HPE; AI Transparency a Risk Issue

HPC Wire - Wed, 11/15/2017 - 19:06

The HPC market update from Hyperion Research (formerly IDC) at the annual SC conference is a business and social “must,” and this year’s presentation at SC17 played to a SRO crowd at a downtown Denver hotel. This writer has attended several of these breakfasts, and looking back at the evolving list of major trends identified by Hyperion reveals how incredibly fast advanced scale computing is changing.

For example, two years ago at SC15 in Austin, our coverage shows there was not a single mention of AI, machine learning or deep learning. This year, AI is the air we breathe, with the market straining to blast past HPDA into machine learning and deep learning to the point where Hyperion devoted significant time this morning to sounding warning calls about the power of AI and how lack of transparency threatens its future.

But before going into that, let’s first look at Hyperion’s industry update.

Earl Joseph, Hyperion CEO, portrayed 2016 showing “some decent growth” for HPC server sales, growing to $11.2 billion worldwide, with strongest growth in the sector comprising systems priced at $250,000 and more. While the first half of 2017 shows an overall market decline of 3.5 percent, to $5.1 billion, Joseph said the first two quarters typically are softer than the second half of the year.

In the vendor race, HPE has a significant lead with a 36.8 percent share, followed by Dell EMC with 20.5 percent, Lenovo at 7.8 percent and IBM at 4.9 percent. Nearly half of global HPC server sales are in North America, and X86 processors are used in nearly 75 percent of HPC servers sold worldwide.

Hyperion has also studied the proliferation of HPC and its economic impact. Bob Sorensen, VP of research and technology, reported that there are nearly 800 HPC sites in the U.S. He also said a recent Hyperion study shows that HPC-reliant US economic sectors contribute almost 55 percent of the GDP to the US economy, encompassing $9.8 trillion and accounting for over 15.2 million jobs.

Joseph said that while the high end of the HPC market has shown the strongest health with a relatively weak lower end, looking ahead, it will be the mid-market that will drive the industry to $14.8 billion by 2021, which Joseph called “fairly healthy growth.”

He also noted the spiky nature of the supercomputing industry, pointing out that three or four exascale-class machines are scheduled to be delivered in 2021 followed by only one or two in subsequent years, resulting in the industry showing “unbelievable growth followed by decline.”

Hyperion also analyses the industry by vertical sectors. The national labs, academia and the Department of Defense comprise nearly half of the total market, with CAE and bio-sciences representing the largest commercial sectors.

From a revenue perspective, AI (broadly defined) offers strong market potential. Senior research VP Steven Conway said Hyperion forecasts a rapid ramp up to $1.2 billion by 2021. But AI – particular its black box nature – is a double edged sword.

But problems of learning machine transparency pose a threat to the future of the nascent AI market.

He explained the “gigantic” difference between deep learning and machine learning. “With machine learning, you train the computer to do some things, and then it does them. With deep learning, you train the computer and then it goes beyond the training and learns on its own.”

He quoted recent warnings by Stephen Hawking that the development of AI “is one of the pivotal moments in human history,” and he cited Dr. Eng Lim Goh of HPE to the effect that “we humans can’t teach computers how to think like we do because we don’t know how humans think. So they’re (computers) think the way they do. That’s a big difference. They can think on their own. And they’re really capable of culture, of learning from each other – that’s been demonstrated.”

While the opacity of deep learning systems raises the dystopic specter of machines moving beyond human control, the more immediate problem of transparency is trust. Citing the famous case earlier this year of Google admitting it didn’t understand how its AlphaGo system defeated the world champion Go player, Conway said it’s one thing to not know how a machine won a board game and another to entrust your life, literally, to a computer – i.e., an autonomous car or AI-driven medical treatment.

“The problem with deep learning today is that in many instances it’s a black box phenomenon, and that needs to change,” Conway said.

Change is on the way.

In Germany, he said, the first national law has been passed that puts restrictions and imposes ethical standards on autonomous vehicles. Based on the work of a 14-person ethics commission the law requires that self-driving cars must allow humans to take over control of the car. Further if its proven that the car was in control when an accident occurred then the automaker is liable; if a person was in control then the driver is liable.

In addition, autonomous cars in Germany will not be programmed to make “demographic” ethical decisions along the lines of letting an elderly person die rather than a baby.

In the U.S., Conway said, limited steps have been taken hold autonomous cars accountable, with the National Highway Transportation Safety Administration issuing a document requiring a “black box” in self-driving vehicles similar to those on airplanes that enable post facto reconstruction of accidents.

“When, as inevitably will happen, we start seeing some self-driving vehicles where they can’t avoid an accident and there’s injury  or death involved, particularly in the early years, then insurance companies, automakers and families will want to know how it happened, why it happened,” Conway said.

AI accountability also is a major issue in the medical profession – if treatments fail that are based on deep learning system recommendations, then questions will be asked. Conway said another key concern with precision medical systems, such as IBM Watson, is that the data selected to train the system can be flawed due to human bias.

Hyperion also announced a mapping project that tracks the nearly 800 HPC sites around the country, along with an initiative to study the emergence of quantum computing. The company also announced the winners of its HPC user and vendor innovation awards.

The post Hyperion Market Update: ‘Decent’ Growth Led by HPE; AI Transparency a Risk Issue appeared first on HPCwire.

Enyx Premieres TCP and UDP Offload Engines for Intel Stratix 10 FPGA On REFLEX CES XpressGXS10-FH200G Board

HPC Wire - Wed, 11/15/2017 - 17:35

DENVER, Nov. 15, 2017 — Enyx, a world-class pioneer in ultra-low latency FPGA-based technology and solutions, is pleased to announce its enterprise-class TCP/IP, UDP/IP and MAC network connectivity Intellectual Property (IP) Cores for FPGAs and SoCs support for the high-performance REFLEX CES XpressGXS10-FH200G PCIe Board, which features a Stratix 10 GX FPGA from Intel’s new top-of-the-line 14nm Stratix 10 family

Enyx network connectivity IP Cores are addressing the growing throughput and hardware acceleration needs of the datacenter industry and are performing network protocol offloading for applications, such as network security enabled NICs, smart NICs, high performance data distribution, custom packet filtering and high bandwidth bridges. Enyx also provides custom project implementation through Enyx Design Services as part of a complete and customized Smart NIC or Smart Switch solution.

“We are pleased to collaborate with our valued partner REFLEX CES to offer the industry-first TCP and UDP full hardware stacks on Intel’s new, cutting-edge Stratix 10 FPGAs,” says Eric Rovira Ricard, VP Business Development North America at Enyx. “Intel is making FPGA technology ready for data centers, opening new areas for hardware offloading applications in high performance computing, and Enyx is proud to provide the most mature and feature rich network protocol stacks for seamless, FPGA-enabled network connectivity on the latest devices.”

“We are delighted to work with Enyx to offer the best-in-class UDP & TCP IP low latency reference design on our Stratix 10 FPGA board first to market for the Finance and Networking applications, and therefore providing a fast and trusted solution,” said Eric Penain, Chief Business Officer at REFLEX CES.

Enyx nxTCP and nxUDP IP Cores feature full RTL Layers 2, 3, 4 implementations with integrated 40G/25G/10G/1G MAC, compliant with the IEEE 802.3 standards, supporting ARP, IPv4, ICMP, IGMP and TCP/UDP protocols. nxTCP and nxUDP are designed to work seamlessly on Intel (formerly Altera) and Xilinx FPGA and SoC designs. Enyx TCP implementation on Intel Stratix 10 GX devices feature latencies of less than 60 ns in transmission and 110 ns in reception and can also manage up to 32,768 TCP sessions in parallel.

REFLEX CES XpressGXS10-FH200G is the first commercially available PCIe board supporting the 14nm Intel Stratix 10 FPGA family. REFLEX CES XpressGXS10-FH200G PCIe board includes the biggest 2800 KLE Stratix 10 density for processing intensive and various data algorithms with its mix of memory capabilities in DDR4 and QDR2+. It has an optical interface capability of 200Gbit via two QSFP28 cages and uses PCIe gen3 x16. An additional 200Gbit board-to-board interface is provided using a firefly connection. The footprint is compatible with SoC FPGA’s enabled HPS access via the Ethernet interface on the PCIe bracket side. REFLEX CES is a certified board partner of Enyx.

Starting in 2018, the Intel Stratix 10 version downloadable package will be available and will include a reference design for the REFLEX CES XpressGXS10-FH200G PCIe board.

Enyx made this announcement today at the SC17 conference in Denver where it is currently presenting its technology product line and services.

About Enyx

Enyx is a leading developer and provider of FPGA-based ultra-low latency technologies and solutions. Enyx Technology & Design Services division provides design services and connectivity IP cores for FPGA and SoC, for tailored Smart NICs and Smart Switches. Enyx Technology & Design Services division has engaged with over 50 customers world-wide, including hedge funds, exchanges, top-tier investment banks, telecom operators, research labs, universities, and technology manufacturers for the defense, military, aeronautics, aerospace and high-performance computing industries.

For more information, visit www.enyx.com


Recognized for its expertise in high-speed applications, analog and hardened systems, REFLEX CES has become a leading partner with major industrial companies. REFLEX CES simplifies the adoption of FPGA technology with its leading-edge FPGA-based custom embedded and complex systems. REFLEX CES FPGA network platforms enable better flexibility and ease of programming, offering a faster and most powerful board, and reducing the customers’ technology risks and time to market. The company provides FPGA COTS boards for several markets, including the Finance market where Ultra Low Latency capability is a key element, and other markets like Networking.

Source: Enyx

The post Enyx Premieres TCP and UDP Offload Engines for Intel Stratix 10 FPGA On REFLEX CES XpressGXS10-FH200G Board appeared first on HPCwire.

CoolIT Systems Selected for Largest Datacenter Deployment in Company History with Non-Disclosed U.S. Customer

HPC Wire - Wed, 11/15/2017 - 10:26

CALGARY, Alberta, Nov. 15, 2017 — CoolIT Systems (CoolIT), global leader in energy efficient liquid cooling for HPC, Cloud and Hyperscale, has been selected to deploy its largest data center solution to date for a non-disclosed U.S. customer, supplied through an OEM partner.

The data center installation requires over 30,000 CoolIT Systems Rack DCLC coldplates with supporting manifolds, providing direct contact liquid cooling to Intel® Xeon Phi Processors.

“The sheer magnitude of this deal reaffirms the global migration towards adopting more energy efficient data center cooling solutions,” said CoolIT Systems CEO, Geoff Lyon. “CoolIT continues to be the leading choice for server manufacturers and we are proud to provide our OEM partner with factory-installed liquid cooling to support this major U.S. installation.”

CoolIT Systems Executives and Technical Experts will be available during SC17 in Denver to discuss warm water liquid cooling solutions for high performance servers and data center environments. Visit booth 1601 from Monday 13th – Thursday 16th November 2017.

About CoolIT Systems

CoolIT Systems, Inc. is the world leader in energy efficient liquid cooling technology for the Data Center, Server and Desktop markets. CoolIT’s Rack DCLC platform is a modular, rack-based, advanced cooling solution that allows for dramatic increases in rack densities, component performance, and power efficiencies. The technology can be deployed with any server and in any rack making it a truly flexible solution. For more information about CoolIT Systems and its technology, visit www.coolitsystems.com.

Source: CoolIT Systems

The post CoolIT Systems Selected for Largest Datacenter Deployment in Company History with Non-Disclosed U.S. Customer appeared first on HPCwire.

Inspur Partners with NEC Deutschland GmbH to Develop European HPC Market

HPC Wire - Wed, 11/15/2017 - 10:20

On November 14 (US Mountain Time), Inspur and NEC Deutschland GmbH announced their strategic cooperation at the 2017 Supercomputing Conference (SC17) in Denver. The two companies will work closely together on products, technologies, and markets to jointly develop the European high-performance computing (HPC) market.

The two parties will focus on the HPC field, primarily relating to European educational and research institutions, as well as commercial and industrial enterprises. By combining their mutual strengths in products, technologies, and market position, Inspur and NEC Deutschland GmbH will jointly provide European customers with trusted HPC product and solutions, helping them to make significant advances in HPC-related areas such as traditional scientific engineering calculations, high-performance data analysis, and emergent AI computing.

“NEC is a prestigious global player and a distinguished HPC solution provider with a broad customer base in the EMEA region and strong R&D and delivery capabilities,” said Jay Zhang, Vice President of Inspur’s overseas headquarters. “This strategic partnership with NEC will accelerate Inspur’s globalization, and with the help of NEC’s excellent market foundation, Inspur will be able to deliver the world’s leading HPC products and services to European customers.

“Inspur is a leading global server manufacturer, with advanced products and extensive experience in the HPC and AI fields, ” said Yuichi Kojima, Managing Director of NEC Deutschland GmbH and Vice President for HPC EMEA. This cooperation between NEC and Inspur will lead to more diverse, competitive HPC products and solutions, and provide European customers with more high-quality choices.”

Sources say that Inspur and NEC Deutschland GmbH have already cooperated on a number of projects and are just entering into the delivery stage of a project providing one of the world’s leading car manufacturers with a jointly designed and constructed high-performance computing cluster.

About Inspur

As one of the Top four server vendors in the world, Inspur has met the highest international standards in system design, R&D and construction for HPC systems. With a strong design and production capability, Inspur supplies solutions covering air/water-cooling infrastructure as well as full line server nodes products of 2-socket, 4-socket, 8-socket, and 32-socket and high-performance storage. It provides adaptive HPC-feature-driven solutions to achieve customized system and applications.

About NEC Deutschland GmbH

NEC Deutschland GmbH is a wholly owned subsidiary of NEC Europe Ltd. and is a leading provider of HPC solutions, focusing on sustained performance for real-life scientific and engineering applications. To achieve this goal NEC delivers technology and professional services to industry and academia. Linux-based HPC clusters as well as our high-end vector systems meet the different needs of different customers in the most flexible way. Energy-efficiency is one of the key design objectives, addressed by advanced cooling technologies or by the high-bandwidth vector-architecture, which delivers unprecedented efficiency on real world code. The service capabilities from the operation of complex systems to the optimization of scientific codes and NEC’s storage-appliances complete our solution offering.

The post Inspur Partners with NEC Deutschland GmbH to Develop European HPC Market appeared first on HPCwire.

One Stop Systems Licenses ION Flash Storage Accelerator Software to Continue Development

HPC Wire - Wed, 11/15/2017 - 09:17

DENVER, Nov. 15, 2017 — One Stop Systems (OSS), a leading provider of high performance all-flash arrays, announced a license agreement reached earlier this year with Western Digital for Western Digital’s ION Flash Storage Accelerator Software. The agreement allows OSS to develop and market the ION software, providing enhancements to a proven flash file system. ION software has provided a stable platform for many military, finance and other pure high speed flash storage applications. ION has a pedigree of ultra-high performance storage for high definition video content delivery, social media platforms, high-speed data recorders, financial transactions, government training systems and many more applications.

The ION software defines flexibility by supporting a variety of modern storage interfaces such as 100Gb Ethernet, 32Gb Fibre Channel and 100Gb InfiniBand. The interfaces support popular protocols such as iSCSI, SRP and FC. ION also features active-standby high availability (HA) with synchronous replication. Multiple data protection options are provided through HA and the support of RAID 0, RAID 1, RAID 5 or RAID 10. When the application calls for pure speed, the ION software supports LUN masking and JBOD mode for scenarios where high IOPS trumps data protection. The ION software also allows for server consolidation by removing storage bottlenecks in virtualized environments. This allows more virtual machines to be provisioned in a virtualized environment per storage system thereby reducing total cost of ownership verses traditional storage or less efficient, open source flash file systems.

“One Stop Systems has built a team of experienced software developers to enhance the ION Software,” said Steve Cooper, CEO of OSS. “The ION software has proven incredibly powerful and works in conjunction with the OSS Data Storage Unit hardware to provide a new level of performance and features for applications such as real-time HPC, analytics, big data and high speed data recording. The software team will extend OSS capabilities, increasing the functionality and value of our hardware by pairing it with proven leading-edge software to serve the demands of our customers.”

“Licensing the ION software to a long-time partner such as One Stop Systems benefits all existing and future ION customers,” said Joan Wrabetz, Vice President of Marketing, Data Center Systems, Western Digital Corporation. “The ION Software contains a wide variety of features that provide performance improvements and latency reduction to today’s demanding ’Fast Data’ applications. Western Digital customers would benefit from the ongoing enhancements enabled by this agreement for proven flash storage accelerator software such as ION.”

About One Stop Systems

One Stop Systems designs and manufactures ultra-dense high performance computing (HPC) systems for deep learning, oil and gas exploration, financial trading, media and entertainment, defense and traditional HPC applications requiring the fastest and most efficient data processing. By utilizing the power of the latest GPU accelerators and NVMe flash cards, our systems stay on the cutting edge of the latest technologies. We have a reputation as innovators in hyperconverged and composable infrastructure solutions using the latest technology and design equipment to operate with the highest efficiency. Now OSS offers these exceptional systems to customers for lease or purchase. OSS continuously works to meet our customers’ greater needs.


Source: One Stop Systems

The post One Stop Systems Licenses ION Flash Storage Accelerator Software to Continue Development appeared first on HPCwire.

GLOBALFOUNDRIES Demonstrates Industry-Leading 112G Technology for Next-Generation Connectivity Solutions

HPC Wire - Wed, 11/15/2017 - 08:59

SANTA CLARA, Calif., Nov. 15, 2017 — GLOBALFOUNDRIES today announced it has demonstrated the next generation of 112Gbps SerDes capability. GF’s High Speed SerDes (HSS) solutions include best-in-class architecture for 112G to 56G, 30G and 16G SerDes IPs to enable connectivity for cloud computing, hyperscale data center, and networking applications.

“GF’s demonstration of 112Gbps SerDes architecture  establishes  the capability  of running extremely high, next-generation interconnect technology that can deliver long-reach capabilities to data center and enterprise applications,” said Mike Cadigan, senior vice president of global sales and business development at GF. “As a result, our customers will soon have access to design best-in-class ASIC solutions to meet the explosive bandwidth growth in data center and networking applications as the industry transitions to a new era of connected intelligence.”

The 112G SerDes is designed in GF’s high-performance ASIC FX-14 technology, and is capable of supporting several multi-level signaling schemes while targeting 25dB+ insertion loss interconnects.  Flexibility has been built into the 112G SerDes platform in order to analyze the efficacy of a variety of higher level encoding schemes such as Forward Error Correction.

The learning from this effort is being used to develop GF’s 112G SerDes IPs in FX-7 technology and leverages the foundry’s leading-edge 7nm FinFET technology platform, 7LP, to enable high-speed connectivity and low-power solutions as well as optical variants, for current and future leading-edge networking, compute and storage applications.

Customers are presently designing advanced ASIC solutions in the 14LPP and 7LP process technologies using the 56Gbps and other FX-14 SerDes cores. GF has demonstrated its ability to deliver a true long-reach SerDes solution by delivering a 56Gbps PAM4 signal on the company’s 14nm FinFET process and is shipping development boards in customer channels.  Enabling 112G communication solutions will allow customers to double chip bandwidth in next generation products.


The post GLOBALFOUNDRIES Demonstrates Industry-Leading 112G Technology for Next-Generation Connectivity Solutions appeared first on HPCwire.

The Betting Window is Closed: Final Student Cluster Competition Betting Odds are in!

HPC Wire - Wed, 11/15/2017 - 08:42

The window has closed and the bettors are clutching their tickets, anxiously awaiting the results of the SC17 Student Cluster Competition. We’ve seen big changes in the betting line over the last few days. In our first odds round-up, NTHU, Tsinghua, and Team Poland were early favorites at around 4 or 5 to 1. But as more (virtual) money has been played, we’ve seen the lines move.

As we’re closing out the pool, the betting favorites are Team Poland at 3.9 to 1 and Team Georgia at 4.6 to 1. Following closely behind are the Asian powerhouse teams NTHU and Tsinghua at 5.6 to 1 and 7.7 to 1 respectively.

There are some experienced teams that probably deserved more action from bettors. Defending SC16 champion USTC, for instance, should definitely not be a 199 to 1 underdog. The same applies to team Nanyang and Illinois – these are both teams on the upswing.

But who can tell what the betting public is thinking? It’s a combination of logic, emotion, and just plain “I gotta hunch” thinking. We’ll see how it all plays out.

I’ll be writing up a comprehensive “Who Won & Why” article in the days following the competition. In this story, we’ll look at the top finishers and compare what really happened to what our bettors thought would happen. If you ask me, there are a lot of bettors who will be tearing up their stubs in anger and crying in their beer when the final results are revealed, but I could be wrong. Stay tuned…..


The post The Betting Window is Closed: Final Student Cluster Competition Betting Odds are in! appeared first on HPCwire.

Engineering of Swedish Quantum Computer Set to Start

HPC Wire - Wed, 11/15/2017 - 08:33

Nov. 15, 2017 — A SEK 1 billion research initiative is setting Sweden on course to a global top position in quantum technology. The focus is on developing a quantum computer with much greater computing power than the best supercomputers of today. The initiative, which is headed up by Professor Per Delsing at Chalmers University of Technology, has been made possible by an anniversary donation of SEK 600 million from the Knut and Alice Wallenberg Foundation.

The progress of research in quantum technology in recent years has brought the world to the brink of a new technology revolution – the second quantum revolution. Researchers have learnt to control individual quantum systems such as individual atoms, electrons and particles of light, which is opening the door to completely new possibilities. Extremely rapid computers, intercept-proof communications and hyper-sensitive measurement methods are in sight.

A major Swedish initiative – the Wallenberg Centre for Quantum Technology – is now being launched under the leadership of Chalmers University of Technology to contribute to, and implement the second quantum revolution. Some 50 researchers are to be recruited under the decade-long research programme which begins in January 2018. In addition to the donation from the Knut and Alice Wallenberg Foundation further funds are coming from industry, Chalmers University of Technology and other universities, resulting in a total budget of close to SEK 1 billion.

The programme’s focus project centres on the engineering of a quantum computer based on superconducting circuits. The smallest building block of the quantum computer – the qubit – is based on principles which are entirely different from those of today’s computers, thus enabling the quantum computer to process vast quantities of data using relatively few qubits.

“Our goal is to have a functioning quantum computer with at least a hundred qubits. Such a computer has far greater computing power than the best supercomputers of today and can be used, for example, to solve optimisation problems, advanced machine learning, and heavy calculations of the properties of molecules,” says Per Delsing, Professor of Quantum Device Physics at Chalmers University of Technology and the initiative’s programme director.

See and hear the researchers tell their story in a video on Youtube: The Quantum Revolution

There is a great deal of interest in quantum technology throughout the world. Major investments are being made in the USA, Canada, Japan and China and the EU is launching a Quantum Technology Flagship in 2019. Companies such as Google and IBM are also investing in quantum computers and, like Chalmers, have chosen to base them on superconducting circuits. Policy-makers and business managers are starting to realise that quantum technology has the potential to change our society significantly, through improved artificial intelligence, secure encryption and more efficient design of medicines and materials.

“If Sweden is to continue to be a leading nation we need to be at the forefront in these fields. By focusing on the long-term expansion of expertise and by attracting the best young researchers we can put Sweden on the quantum technology map in the long term. There are no shortcuts. By investing in basic research we can ensure that the necessary infrastructure is in place so that over time other players and companies can take over and develop applications and new technologies,” says Peter Wallenberg Jr, chairman of the Knut and Alice Wallenberg Foundation.

In addition to the focus project the research programme includes a national excellence initiative with the aim of carrying out research and building up expertise in the four sub-areas of quantum technology: quantum computers, quantum simulators, quantum communication and quantum sensors. Chalmers University of Technology is coordinating the first two sub-areas. The expansion of expertise in quantum communication is headed up by Professor Gunnar Björk at KTH Royal Institute of Technology, and Professor Stefan Kröll at Lund University is coordinating the quantum sensor field.

Chalmers researchers have been working on superconducting qubits for almost 20 years and have made many contributions to enhance knowledge in the field, including publications in Nature and Science. They were among the first in the world to create a superconducting qubit, and have explored a completely new area of physics through wide-ranging experiments on individual qubits.

“I am pleased that our quantum physics researchers, along with colleagues in the rest of Sweden, will have this opportunity to focus on a specific and important goal in a way that all of Sweden can benefit from the knowledge acquired. I would also like to extend my warmest thanks to the Wallenberg Foundation for its deep commitment and long-term support,” says Stefan Bengtsson, President and CEO of Chalmers.

In parallel with this, the Knut and Alice Wallenberg Foundation is investing SEK 1 billion in artificial intelligence, channelled through the Wallenberg Autonomous Systems and Software Program (WASP), which was launched in 2015. Details of the investment can be found in a press release from Linköping University: Billion-investment in Swedish AI research

Read more about the link between the two anniversary donations in a press release from the Knut and Alice Wallenberg Foundation: SEK 1.6 billion for artificial intelligence and quantum technology

Wallenberg Centre for Quantum Technology

  • Wallenberg Centre for Quantum Technology is a ten-year SEK 1 billion initiative aimed at bringing Swedish research and industry to the front of the second quantum revolution.
  • The research programme aims to develop and secure Swedish competence in all areas of quantum technology: quantum computing, quantum simulation, quantum communications and quantum sensing.
  • The research programme includes a focus project aimed at developing a quantum computer and an excellence programme covering the four areas of quantum technology.
  • Wallenberg Centre for Quantum Technology is led by, and is largely located at Chalmers University of Technology. The areas of quantum communication and quantum sensors are coordinated by KTH Royal Institute of Technology and Lund University.
  • The initiative includes a graduate research school, a postdoctoral program, a guest researcher programme and funds for recruiting young researchers. It will ensure Swedish long-term expertise in quantum technology, even after the end of the programme.
  • Collaboration with several industry partners ensures that the areas of application become relevant to Swedish industry.

Read more in the programme fact sheet: Wallenberg Centre for Quantum Technology

The second quantum revolution

In the 20th century, the first quantum revolution took place. It gave us inventions like the laser and transistor – inventions that underlie the entire information technology that forms today’s society.

After many years of basic research on strange quantum phenomena such as superposition, entanglement and squeezed states, scientists have learned to control individual quantum systems as individual atoms, electrons and photons. The world record currently stands at 20 qubits, but rapid progress is being made each month. Applications such as extremely fast quantum devices, intercept-proof communications and hyper-sensitive measuring methods are in sight.

Therefore, heavy investments in quantum technology are being made throughout the world. The EU launches a ten-year venture of one billion euros in 2019. Even larger programmes exist in North America, Asia and Australia. IT companies like Google, IBM, Intel and Microsoft are also making significant investments. Safe and fast communication is a strong driving force for quantum technology. Already today there are commercial systems that can transmit quantum encryption keys through an unbroken optical fibre over 100 kilometres, although at a relatively low speed.

An imminent milestone that scientists are struggling to achieve is to demonstrate quantum supremacy, which means solving a problem beyond reach even for the most powerful future classic computer. This requires at least 50 qubits. This will be done by means of a quantum simulator, a simpler form of quantum computer. Useful applications of quantum simulation are expected within five years. Realizing a functioning programmable quantum computer will take significantly longer.

Mankind’s knowledge about the world and our technical advances are limited by what we can measure, and how accurately. Researchers are also learning to use individual particles, such as photons and electrons, as sensors in measurements of force, gravity, electrical fields, etc. With quantum technology, the measuring power is pushed far beyond what was previously possible.

Read more about central quantum phenomena in the fact sheet: Quantum technology

Source: Chalmers University of Technology

The post Engineering of Swedish Quantum Computer Set to Start appeared first on HPCwire.

2017 Student Cluster Competition Benchmarks, Workloads, and Pre-Planned Disasters

HPC Wire - Wed, 11/15/2017 - 08:21

The students competing in the 2017 Student Cluster Competition in Denver are facing a grueling 48 hour marathon of HPC benchmarks and real scientific applications. There are also a few surprised in store for them. Let’s take a look at what they’ll have to do in order to finish in the winner’s circle (although there isn’t really a circle for them to stand in).

First up on Monday, the teams had to run HPCG, which stresses the systems in brutal fashion. Jack Dongarra’s newest benchmark creation is designed to much more closely mimic how HPC systems are used today and has quickly become a staple for Student Cluster Competitions worldwide.

After the teams submit their best HPCG score, it’s on to the venerable LINPACK. Teams will configure their cluster in its final form for this run and vie to submit the highest score possible. After completing LINPACK, the teams await the Monday evening 7:30 mark when they’ll get the data sets for the scientific applications and the real work begins.

Scientific applications this year include:

LAAMPS: LAAMPS will be used for the “Reproducibility Challenge” where student teams are tasked with reproducing the results from a research paper that was published 2016. The students will have to prove (or disprove) the results in the paper and justify their position. You can read more about this task and the others here .

MrBayes: will be used to help figure out how to reduce the ravages of Cassava Brown Streak virus and Uganda Brown Streak virus on the African food staple cassava. The viruses are transmitted by white flies, and the specific genome of the white fly is under study. In this task, students are working to figure out how to best protect the cassava plant from disease.

Born: is a seismic imaging tool that is used in industry to understand what’s underneath all of the grass, dirt, trees and other stuff on top of the earth. Born can help identify reservoirs of gas and liquids under the earth, both the position and actual size of the deposit. Students will take data sets and propagate the data both forward and backward in time – although I’m pretty sure there isn’t any actual time travel involved in this exercise.

Mystery Application: It is M-Pass Atmosphere, which will be used to study what would happen if we were able to sequester excess CO2 in Antarctica. How would CO2 flow in the norther hemisphere? What would that look like? Students will be coming up with the answers.

New this year is a cloud component to the competition. Students will be given a budget to use with Cycle Computing on the Microsoft Azure Cloud. The teams will have the choice to run their workloads on premise as usual, or to run them on the Azure cloud. However, nothing in the Student Cluster Competition is free. Students have a limited cloud budget, only $600, and there are harsh penalties for running over their budgets.

Another twist in this year’s competition is the “Power Shutoff Activity” where the power to all of the student booths will be shut off. This recreates what happened at the first cluster competition in 2007 when the Reno convention center power failed.

We don’t know exactly how many power failures the students will have to endure. As the organizers said “it’s a number between zero and infinity”. But I can tell you that power was dropped at 5:00 am Denver time Tuesday morning. The teams that had people staffing their booth were fine, while a handful of teams lost some cluster time as their systems remained in a power off state until they showed up at the convention center. Fun times.

Next up we’ll be looking at the team configurations, the final betting odds, and start getting into our highly anticipated video interviews. Stay tuned……

The post 2017 Student Cluster Competition Benchmarks, Workloads, and Pre-Planned Disasters appeared first on HPCwire.

Netcope Unveils Cloud-Based Solution to Make FPGA Programming More Accessible

HPC Wire - Wed, 11/15/2017 - 08:07

Nov. 15, 2017 — Netcope Technologies has recently unveiled a product going by the moniker NP4 (Netcope P4). With it, the company has brought to daylight an entirely new concept of “Firmware as a Service”. With it, FPGA firmware design just became completely outsourceable to anyone who can describe network functions with P4 language.

The P4 programming language is a high level declarative language, so it is much more accessible to network operators who wish to leverage programmable hardware in their solutions, but lack HDL skills. The spread of P4 in the networking industry is being spearheaded by Barefoot Networks and Netcope, as a proud member of P4 Consortium, started to work with P4 immediately after it had been made available to the broader public. P4 to VHDL translator, Netcope’s first P4-related technology, has been announced roughly a year ago.

But the P4 to VHDL translator was just a sneak peek of what Netcope planned to do with P4. Since all of the company’s products are based around FPGA technology, the result expectedly took form of full-on P4+FPGA combo. It is the cloud-based service NP4 that drastically simplifies the process of FPGA programming. Write the P4 code, upload it to the NP4 cloud, wait for the application to autonomously translate the P4 code into a VHDL code and do the firmware synthesis. Then download the firmware bitstream and upload it to your FPGA appliance. That is the process of using NP4 as a customer.

From a broader perspective, NP4 is a representation of the tangible change that the network industry is going to be affected by. It is the “Firmware as a Service” concept. The process of firmware synthesis outsourcing described above can be taken advantage of in 2 major ways. The first way is relevant to those vendors and network operators who don’t have any HDL engineers in their teams and wish to enjoy the flexibility of programmable hardware. The second way is relevant to those who are HDL proficient and prefer to have total control over their solutions, but wish to outsource some parts of the process in order to increase the overall output of their whole R&D operation.

Either way, just about anyone with programmable hardware in their networks can benefit from this new method of reconfiguration of network functions. “In addition to reducing time to market, NP4 brings additional diferentiation to the world of FPGA accelerators. Vendors of FPGA-accelerated appliances can now easily offer completely different feature sets, even if they use very similar hardware platform. We are fully committed to continuously improve the NP4 cloud service, and the list of features and extensions that we plan to add is impressive. In addition, everything in NP4 is designed for throughput of 100 Gbps or more.” says Viktor Puš, CTO of Netcope Technologies.

Source: Netcope Technologies

The post Netcope Unveils Cloud-Based Solution to Make FPGA Programming More Accessible appeared first on HPCwire.

TYAN and Nyriad Announce First GPU-Accelerated Storage Server Solution for Ultra-Dense Storage

HPC Wire - Wed, 11/15/2017 - 07:39

DENVER, Nov. 15, 2017 — TYAN, in partnership with Nyriad, announces a revolutionary advance in ultra-dense GPU-accelerated storage to be demonstrated at Supercomputing 2017 (SC17).

TYAN has qualified the Thunder SX (FA100-B7118) storage server with an NVIDIA Pascal GPU to enable Nyriad’s NSULATE™ GPU-accelerated storage. NSULATE uses a GPU to perform erasure coding and cryptographic checksumming. This enables NSULATE to support hundreds of device failures within a large storage array.

This solution overcomes the limitations of traditional RAID-based storage and enables low cost, high-density data storage. It supports the creation of DAS arrays with hundreds of parallel storage devices, able to provide high performance even with massive array degradation. The specific hardware implementation also allows GPU computing on the storage nodes, resulting in minimal data movement and saving on network fabric.

Matthew Simmons, CEO of Nyriad Ltd., said, “We are thrilled to have TYAN as a qualified hardware partner for NSULATE as getting speed, storage density, and extreme error resilience has always been a matter of making difficult tradeoffs for each application. Using the GPU for storage processing has finally made it possible to have all of these features in a single solution, and now TYAN has it in their Thunder SX (FA100-B7118) storage server.”

TYAN expressed its excitement about the partnership with Nyriad as GPU computing directly on storage servers will enable it to offer server systems with lower power consumption and minimal data movement.

Danny Hsu, Vice President of MiTAC Computing Technology Corporation’s TYAN Business Unit stated, “The continued growth of HPC applications has enabled solutions for science, medicine and many other fields of study, while at the same time further increased the need for faster, more resilient storage and compute requirements. Our partnership with Nyriad has yielded an incredible solution combining GPU compute and high speed storage, based on the Thunder SX FA100-B7118 server.”

TYAN and Nyriad will demonstrate the solution at Supercomputing 2017 in Denver, Colorado, on November 13-16, Booth #1269.

About TYAN

TYAN, as a leading server brand of MiTAC Computing Technology Corporation under the MiTAC Group (TSE:3706), designs, manufactures and markets advanced x86 and x86-64 server/workstation board technology, platforms and server solution products. Its products are sold to OEMs, VARs, System Integrators and Resellers worldwide for a wide range of applications. TYAN enables its customers to be technology leaders by providing scalable, highly-integrated, and reliable products for a wide range of applications such as server appliances and solutions for HPC, hyper-scale/data center, server storage and security appliance markets. For more information, visit MiTAC’s website at http://www.mic-holdings.com or TYAN’s website at http://www.tyan.com.

About Nyriad

Nyriad is an exascale computing company specialising in advanced data storage solutions for big data and high performance computing. Officially the first commercial spin-out from the Square Kilometre Array Project, the company was forced to rethink the relationship between storage, processing and bandwidth to achieve a breakthrough in system stability and performance capable of processing and storing over 160Tb/s of radio antennae data in real-time, within a power budget impossible with any modern IT solutions. For more information, visit http://www.nyriad.com.

Source: Nyriad

The post TYAN and Nyriad Announce First GPU-Accelerated Storage Server Solution for Ultra-Dense Storage appeared first on HPCwire.

Nvidia Focuses Its Cloud Containers on HPC Applications

HPC Wire - Tue, 11/14/2017 - 21:30

Having migrated its top-of-the-line datacenter GPU to the largest cloud vendors, Nvidia is touting its Volta architecture for a range of scientific computing tasks as well as AI development with the addition of a container registry designed to deploy its GPU cloud for everything from visualization to drug discovery.

Nvidia CEO Jensen Huang, SC17 in Denver

In its drive to expand access to its Volta architecture, Nvidia announced the availability of its Tesla V100 GPU on Microsoft Azure. Azure is the latest cloud service to join the chipmaker’s growing list of public and private cloud services providers along server makers. Most are offering the “GPU accelerated” services for AI development projects such as training deep learning models that require more processing cores and access to big data.

Moving beyond the AI market, Nvidia on Monday (Nov. 13) unveiled a container registry designed to ease deployment of HPC applications in the cloud. The container registry for scientific computing applications and visualization tools would connect researchers with most GPU-optimized HPC software, the company said during this week’s SC17 conference in Denver.

Last month, the company introduced deep learning applications and AI frameworks in its Nvidia GPU Cloud (NGC) container registry. The AI container registry was rolled out on Amazon Web Services’ Elastic Compute Cloud instances running on Tesla V100 GPUs.

The HPC application containers announced this week include a long list of third-party scientific applications. HPC visualization containers are available in beta on the GPU cloud.

As GPU processing moves wholesale to the cloud and datacenters, easing application deployment was the next logical step as Nvidia extends its reach beyond AI development to scientific computing. (The company notes that the 2017 Nobel Prize winners in chemistry and physics used it CUDA parallel computing and API model. Nvidia’s Volta architecture includes more than 5,000 CUDA cores.)

HPC containers are designed to package the libraries and dependencies needed to run scientific applications on top of container infrastructure such as Docker Engine. The cloud container registry for delivering HPC applications uses Nvidia’s Docker distribution to run visualizations and other tasks in GPU-accelerated clouds. The service is available now.

Underpinning these scientific workloads in the cloud is the Volta architecture, asserts Nvidia CEO Jensen Huang. “Volta has now enabled every researcher in the world to access…the most advanced high-performance computer in the world at the lowest possible price,” Huang claimed during SC17. “You can rent yourself a supercomputer for three dollars” per hour.

The other part of the GPU equation is the software stack and how it remains optimized. Hence, Nvidia has placed software components in the GPU cloud via its container registry. The containerized software stack can then be downloaded from Nvidia’s cloud and datacenter partners.

Emphasizing Nvidia’s drive to make GPU processing more accessible, Huang concluded: “In the final analysis, it’s got to be simple.”

Nvidia also took advantage of the SC17 launch pad to announce it is building a new supercomputer to enable high-performance workflows inside its company. SaturnV with Volta continues the tradition of the DGX-1 SaturnV that the company announced last year at SC16, but swaps out the Pascal-based P100s with Volta-based V100s. Nvidia is also greatly expanding the system: from 124 nodes to 660 nodes. Once complete in early 2018, it will offer 40 petaflops of peak double-precision floating point performance, Nvidia said. An early version of the system appeared on the 50th Top500 list (revealed Monday), delivering 1.07 Linpack petaflops in 30 DGX-1 nodes, sufficient for a 149th ranking. That system, installed at Nvidia headquarters in San Jose, Calif., also secured the fourth highest spot on the Green500 listing.

“This is one of the fastest and greenest supercomputers in the world and we use it for our high-performance computing software stack development,” said Huang.

“I believe this is the future of software development,” he continued. “Until now, most of our software engineers coded on their laptop, they compiled it and ran regression tests in the datacenter. Now we have to have our own supercomputing infrastructure. I believe every single industry, every single company will eventually have to have high performance computing infrastructures, opening up the opportunity for the HPC industry.”

–Tiffany Trader contributed to this report.

The post Nvidia Focuses Its Cloud Containers on HPC Applications appeared first on HPCwire.

SkyScale Announces Cloud Solutions Partnership with Rescale and Availability of NVIDIA’s Tesla V100 GPUs on its Cloud Platform

HPC Wire - Tue, 11/14/2017 - 10:44

DENVER, Nov. 14, 2017 — SkyScale, a leader in accelerated cloud solutions, today simultaneously announced its partnership with Rescale, the foremost cloud simulation and HPC solution provider, and availability of NVIDIA Tesla V100 GPU accelerators on its GPU as a Service cloud platform. Enhancing SkyScale’s dedicated, cloud-based ultra-fast multi-GPU platforms for deep learning and HPC applications, the Tesla V100 offers the performance of 100 CPUs in a single GPU—enabling data scientists, researchers, and engineers to tackle the most difficult deep learning and HPC challenges.

SkyScale’s V100 offerings are dedicated bare-metal systems with four, eight, or sixteen V100 GPUs in a single node, which can also be clustered. Interconnect technologies include PCIe and NVIDIA NVLink™ high-speed interconnects, running up to 300 GB/s; five times the bandwidth of PCIe.

“There has been a lot of anticipation for the V100 processor, and justifiably so,” said Tim Miller, president of SkyScale. “The performance of the NVIDIA V100 with NVLink™ truly represents a leap forward over previous generations, particularly for AI and HPC. SkyScale is proud to be among the very first to offer such advanced GPU performance as a service in dedicated systems, which are generally faster and more secure than virtualized machines.”

Partnership with Rescale

“Customers can access our systems a couple of different ways,” remarked Jim Ison, V.P. of Sales for SkyScale. “They can rent our ever-expanding, fully-provisioned and dedicated systems for a week or longer through our website, www.SkyScale.com, and these same systems are now accessible by the hour through our on-demand partnership with Rescale (www.Rescale.com).”

“We are excited to have added SkyScale’s dedicated, ultra-fast V100 and P100 multi-GPU nodes to our ScaleX platform,” said Gabriel Broner, VP and GM of HPC at Rescale. “In turn, our partnership brings SkyScale the benefits of on-demand elasticity and access flexibility to their non-virtualized systems, along with industry-leading software applications for deep learning and scientific and engineering simulations.”

For more information on SkyScale’s GPU as a Service, please visit www.SkyScale.com or call (888) Visit SkyScale at SC17 in Denver, CO from November 13th-16th in Booth #2049

About SkyScale

SkyScale is a world-class provider of cloud-based, ultra-fast multi-GPU platforms for HPC and the fastest deep learning performance available as a service anywhere in the world, and hosts cutting-edge and highly secure private clouds: all in industry-leading datacenters featuring unmatched reliability and physical and cyber security. Related services include solution consulting, simulation and tuning. SkyScale is an NVIDIA Authorized Cloud Service Provider. Visit www.SkyScale.com.

About Rescale

Rescale is the global leader for high-performance computing simulations and deep learning in the cloud. Trusted by the Global Fortune 500, Rescale empowers the world’s top scientists and engineers to develop the most innovative new products and perform groundbreaking research and development faster and at lower cost. Rescale’s ScaleX platform transforms traditional fixed IT resources into flexible hybrid, private, and public cloud resources—built on the largest and most powerful high-performance computing network in the world. For more information on Rescale’s ScaleX platform, visit http://www.rescale.com.

Source: SkyScale

The post SkyScale Announces Cloud Solutions Partnership with Rescale and Availability of NVIDIA’s Tesla V100 GPUs on its Cloud Platform appeared first on HPCwire.

Nyriad Names HPC Systems as Japanese Distributor for GPU Storage Solution NSULATE at SC17

HPC Wire - Tue, 11/14/2017 - 10:37

DENVER, Nov. 14, 2017 — Exascale computing company Nyriad today named HPC Systems as distribution partner for NSULATE to enable GPU-based storage solutions in Japan.

HPC Systems Inc. is a leading system integrator of High Performance Computing (HPC). The company has quickly established itself as a technology and performance leader and is now expanding its product offering to include GPU-accelerated storage solutions.

Nyriad’s NSULATE software enables GPUs to function as storage controllers, replacing RAID controllers and enabling fast, extremely resilient, low power HPC storage solutions.

HPC Systems President Teppei Ono said, “For years we have had to deal with the problems and limitations of RAID. Now we don’t have to make a tradeoff between performance, integrity, and cost. With NSULATE we can have all three, by swapping the RAID card for a GPU.”

NSULATE enables a GPU to accelerate storage processing, providing functions traditionally performed by CPUs and RAID controllers at high speeds in GPU memory, performing cryptographic checksums and high-parity erasure coding, enabling instant recovery from hundreds of simultaneous device failures, which is key for exascale storage.

Nyriad Chief Executive and Co-Founder Matthew Simmons stated, “We are thrilled to have HPC Systems as our distribution partner in Japan. They are respected leaders in the field, so it was a logical choice for us.”

About Nyriad

Nyriad is a New Zealand-based exascale computing company specialising in advanced data storage solutions for big data and high performance computing. Born out of its consulting work on the Square Kilometre Array Project, the company was forced to rethink the relationship between storage, processing and bandwidth to achieve a breakthrough in system stability and performance capable of processing and storing over 160Tb/s of radio antennae data in real-time, within a power budget impossible with any modern IT solutions.

About HPC Systems

HPC Systems Inc. is a leading system integrator of High Performance Computing (HPC) solutions. Since its inception in 2006, the company has quickly established itself as a technology and performance leader in Japanese small-to-mid-range HPC market. The company plans for further growth and developments in world class HPC Cloud solutions, to support our customers research and technological development worldwide.

Source: Nyriad

The post Nyriad Names HPC Systems as Japanese Distributor for GPU Storage Solution NSULATE at SC17 appeared first on HPCwire.

Sandia, Boston University Win Award for Using Machine Learning to Detect Issues

HPC Wire - Tue, 11/14/2017 - 10:32

ALBUQUERQUE, N.M., Nov. 14, 2017 — A team of computer scientists and engineers from Sandia National Laboratories and Boston University recently received a prestigious award at the International Supercomputing conference for their paper on automatically diagnosing problems in supercomputers.

The research, which is in the early stages, could lead to real-time diagnoses that would inform supercomputer operators of any problems and could even autonomously fix the issues, said Jim Brandt, a Sandia computer scientist and author on the paper.

Supercomputers are used for everything from forecasting the weather and cancer research to ensuring U.S. nuclear weapons are safe and reliable without underground testing. As supercomputers get more complex, more interconnected parts and processes can go wrong, said Brandt.

Physical parts can break, previous programs could leave “zombie processes” running that gum up the works, network traffic can cause a bottleneck or a computer code revision could cause issues. These kinds of problems can lead to programs not running to completion and ultimately wasted supercomputer time, Brandt added.

Selecting artificial anomalies and monitoring metrics

Brandt and Vitus Leung, another Sandia computer scientist and paper author, came up with a suite of issues they have encountered in their years of supercomputing experience. Together with researchers from Boston University, they wrote code to re-create the problems or anomalies. Then they ran a variety of programs with and without the anomaly codes on two supercomputers — one at Sandia and a public cloud system that Boston University helps operate.

While the programs were running, the researchers collected lots of data on the process. They monitored how much energy, processor power and memory was being used by each node. Monitoring more than 700 criteria each second with Sandia’s high-performance monitoring system uses less than 0.005 percent of the processing power of Sandia’s supercomputer. The cloud system monitored fewer criteria less frequently but still generated lots of data.

With the vast amounts of monitoring data that can be collected from current supercomputers, it’s hard for a person to look at it and pinpoint the warning signs of a particular issue. However, this is exactly where machine learning excels, said Leung.

Training a supercomputer to diagnose itself

Machine learning is a broad collection of computer algorithms that can find patterns without being explicitly programmed on the important features. The team trained several machine learning algorithms to detect anomalies by comparing data from normal program runs and those with anomalies.

Then they tested the trained algorithms to determine which technique was best at diagnosing the anomalies. One technique, called Random Forest, was particularly adept at analyzing vast quantities of monitoring data, deciding which metrics were important, then determining if the supercomputer was being affected by an anomaly.

To speed up the analysis process, the team calculated various statistics for each metric. Statistical values, such as the average, fifth percentile and 95th percentile, as well as more complex measures of noisiness, trends over time and symmetry, help suggest abnormal behavior and thus potential warning signs. Calculating these values doesn’t take much computer power and they helped streamline the rest of the analysis.

Once the machine learning algorithm is trained, it uses less than 1 percent of the system’s processing power to analyze the data and detect issues.

“I am not an expert in machine learning, I’m just using it as a tool. I’m more interested in figuring out how to take monitoring data to detect problems with the machine. I hope to collaborate with some machine learning experts here at Sandia as we continue to work on this problem,” said Leung.

Leung said the team is continuing this work with more artificial anomalies and more useful programs. Other future work includes validating the diagnostic techniques on real anomalies discovered during normal runs, said Brandt.

Due to the low computational cost of running the machine learning algorithm these diagnostics could be used in real time, which also will need to be tested. Brandt hopes that someday these diagnostics could inform users and system operation staff of anomalies as they occur or even autonomously take action to fix or work around the issue.

This work was funded by National Nuclear Security Administration’s Advanced Simulation and Computing and Department of Energy’s Scientific Discovery through Advanced Computing programs.

About Sandia National Laboratories

Sandia National Laboratories is a multimission laboratory operated by National Technology and Engineering Solutions of Sandia LLC, a wholly owned subsidiary of Honeywell International Inc., for the U.S. Department of Energy’s National Nuclear Security Administration. Sandia Labs has major research and development responsibilities in nuclear deterrence, global security, defense, energy technologies and economic competitiveness, with main facilities in Albuquerque, New Mexico, and Livermore, California.

Source: Sandia National Laboratories

The post Sandia, Boston University Win Award for Using Machine Learning to Detect Issues appeared first on HPCwire.

Aldec to Showcase FPGA-Based Algorithm Accelerators at SC17

HPC Wire - Tue, 11/14/2017 - 10:14

DENVER, Nov. 14, 2017 — Aldec Inc., unveils FPGA-based accelerators, HES-HPC at SC17 to be held on November 12-16, 2017 in Denver Colorado.  Utilizing the latest and most powerful Xilinx UltraScale and UltraScale+ devices with the highest DSP slices, the accelerators are ideal for on premise environments running low-latency applications and algorithms used in Genome Short Reads Alignment, Cloud Computing, Data Mining, High-Frequency Trading and Encryption Code Breaking.

The scalable HES-HPC platforms support PCIe x8 / PCIe x16, include on-board memories, and available in multiple configurations including:

  • 1U or 2U rack built with 7x low profile, PCIe x16 cards with UltraScale+ XCVU9P (~2.5M Logic Cells, 6,840 DSPs), 3x QDR-II+ 144Mb memory, 2x QSFP28 connectors up to 131 Gb/s
  • Desktop System with 1x UltraScale Kintex XCKU115 (~1.5M Logic Cells, 5,520 DSPs) connected to 1x Zynq XC7Z100 as host,  2x 16GB DDR4, 4x 576Mb RLDRAM-3, 2x QSFP+ connectors, USB3.0
  • Single Board with 1x UltraScale Virtex XCVU440 connected to  1x Zynq XC7Z100 as host, 2x 16GB DDR4 and 2x 576Mb RLDRAM-3 memories, 4x FMC and 1x QSFP+ connectors.

“In order to ease the host-to-FPGA transmission, we provide the PCIe-to-AXI hardware infrastructure and C/C++ software API with link speed of 2+GB/s, and it can be used when writing high-level applications for Linux or Windows.” said Louie De Luna, Director of Marketing. “We also provide the Hes.Asic.Proto software package with necessary drivers and utilities for programming and communication with the board.”

Aldec Demonstrations @ Booth#254:

  • Design Encryption Standard (DES) Code Breaker – a reference design for DES Brute Force Code Breaker with a total key combination of 2^56, max input clock frequency 175MHz and 175M combinations/sec for 1 DES instance. The total hack time for 6144 DES instances is ~20h using HES-HPC with 6 Xilinx UltraScale US-440 chips.
  • Vibe Motion Detection – a reference design based on ViBe Background Subtraction algorithm and HES-HPC FPGA-based Accelerator running @1920×1080, 30fps. The image processing background subtraction techniques are utilized to transform and detect moving objects in recorded video. HES-HPC platform provides performance enhancement by utilizing extreme parallel processing capabilities of FPGAs to execute computationally intensive image transformations.
  • Genome Short Reads Alignment – a high-performance reconfigurable FPGA accelerator engine for Renelife ReneGENE, offered on HES-HPC for accurate, and ultra-fast big data mapping and alignment of DNA short-reads from the Next Generation Sequencing (NGS) platforms. AccuRA demonstrates a speedup of over ~ 1500+ x compared to standard heuristic aligners in the market like BFAST which was run on an 8-core 3.5 GHz AMD FX processor with a system memory of 16 GB.

About Aldec

Aldec Inc., headquartered in Henderson, Nevada, is an industry leader in Electronic Design Verification and offers a patented technology suite including: RTL Design, RTL Simulators, Hardware-Assisted Verification, SoC and ASIC Prototyping, Design Rule Checking, IP Cores, Embedded, Requirements Lifecycle Management, DO-254 Functional Verification and Military/Aerospace solutions. www.aldec.com

Source: Aldec

The post Aldec to Showcase FPGA-Based Algorithm Accelerators at SC17 appeared first on HPCwire.

Colorado Engineering, Inc. Unveils First Dual Intel Stratix 10 FPGA at SC17

HPC Wire - Tue, 11/14/2017 - 09:31

COLORADO SPRINGS, Colo., Nov. 14, 2017 — With a legacy of delivering leading-edge custom engineering innovations for industry and military applications, Colorado Engineering, Inc. (CEI), unveiled at Super Computing 2017 in Denver, CO (Booth 1976) the industry’s fastest FPGA Accelerator Card—the WARP II eXtreme High Performance Compute Node (Figure 1).

With its dual Intel Stratix 10 FPGAs, the WARP II PCIe card delivers up to 136GB DDR4 per FPGA, 100GbE, and programmable OpenCL support allowing it to address today’s most challenging data-intense computing problems found in high-density datacenter and cloud service applications.

A Powerful Real-World Solution 

“WARP II represents the most advanced off-the-shelf, PCIe, FPGA compute acceleration card in production,” says Larry Scally, PhD., President and Chief Technology Officer for CEI. “It delivers the power required to turn big data into actionable intelligence. High performance computing applications like machine learning and cognitive computing can all benefit from the performance gains our WARP II delivers.”

Performance Features Abound 

In addition to the dual implementation of Intel’s Stratix® 10 FPGAs, the WARP II supports up to 136GB DDR4 per FPGA and 20 TFLOPS of peak performance (10 TFLOPS per FPGA), making it the most efficient, high performance acceleration card in CEI’s long history.

The WARP II features a myriad of technological advances, including:

  •     4x RDIMM – 264GB Max.
  •     QSFP+ 40/100GbE
  •     PEX PCIe x16 Gen 3
  •     139GB/s FPGA-to-FPGA
  •     Advanced SolidWorks Heatsink Design
  •     8GB Discrete DDR4 RAM
  •     Intel Max 10 FPGA
  •     Freescale K61 Microcontroller
  •     GPU-sized PCIe Form Factor

Designed with a Holistic Approach to Engineering
CEI has long embraced a philosophy of cross-training their engineers. In this way engineers do not simply focus on digital design, RF design, software or firmware design—they focus on the entire process and deliverable.

“Our engineers are involved in all disciplines of engineering,” summarizes Scally. “As a result, the way they treat cause, size, weight, power and performance proves far superior to a product developed by someone with say a software emphasis who will produce a product that is a software solution. We don’t have that problem. Our engineers optimize design by taking this unique holistic approach to product development. The results speak for themselves!”

CEI’s WARP II is currently available and can be ordered directly from CEI at (719) 388-8582 or sales(at)coloradoengineering(dot)com.

About Colorado Engineering, Inc. 

Founded in 2003 with engineering and production facilities located in Colorado Springs, Colorado, Colorado Engineering, Inc. (CEI) supplies Commercial-Off-The-Shelf (COTS) high performance computing hardware and software, as well as tailored solutions, working directly with government agencies and for commercial prime contractors, offering quick turn, innovative solutions with lower cost and higher quality while minimizing risk. CEI leads or partners on new radar systems, software and hardware projects, including leading edge efforts in MOSA applications, reconfigurable RF, THz EM propagation modeling, and associated remote sensing system development.

Colorado Engineering Inc. is certified as a women’s business enterprise by the Women’s Business Enterprise National Council (WBENC), the nation’s largest third-party certifier of the businesses owned and operated by women in the U.S.

A Platinum member of Intel’s FPGA Design Solutions Network, CEI can deliver custom hardware, software, sensor, or mechanical design support for the most demanding computing challenges.

Source: Colorado Engineering, Inc.

The post Colorado Engineering, Inc. Unveils First Dual Intel Stratix 10 FPGA at SC17 appeared first on HPCwire.

One Stop Systems Announces Composable Infrastructure Solutions at SC17

HPC Wire - Tue, 11/14/2017 - 09:06

DENVER, Nov. 14, 2017 — One Stop Systems (OSS), a leading provider of ultra-dense high performance computing (HPC) systems for a multitude of HPC applications, will exhibit several HPC composable infrastructure solutions with partners at SC17 today. The solutions are ideal for data scientists requiring flexible HPC infrastructure across multiple nodes. By adding NVIDIA GPU and NVMe expansion, customers can add unlimited flexibility to their HPC architecture by decoupling the latest innovations in CPU capabilities, NVIDIA GPU performance and NVMe storage into a system called “composable HPC infrastructure.”

“OSS continues to provide the newest solutions to our customers and composable infrastructure is the latest and greatest,” said Steve Cooper, CEO of OSS. “Composable infrastructure using expansion systems allows large numbers of NVIDIA GPUs on the same PCIe or network fabric for use by any node in the datacenter. This flexibility is invaluable for AI, deep learning, RTM, Monte Carlo and image processing applications that benefit from peer-to-peer communication with moderate CPU interaction. Servers, GPUs and storage upgrade on different schedules from the various vendors so composable infrastructure decouples HPC components allowing upgrading at different times, spreading the capital expenditures over many fiscal periods.”

“OSS recently announced new systems that use NVIDIA Tesla V100 GPUs, the most advanced data center GPUs ever built to accelerate AI, HPC, and graphics,” said Paresh Kharya, Group Product Marketing Manager at NVIDIA. “One Stop Systems’ customers can now harness the power of our Volta architecture in their composable infrastructure systems.”

Composable infrastructure allows customers to utilize any number of CPU nodes to dynamically map the optimum number of NVIDIA® Tesla® V100 GPU accelerators and NVMe storage resources to each node required to complete a specific task. When the task completes, the resources return to the cluster pool so they can be mapped to the next set of nodes to run the next task. The composable infrastructure demos in the OSS booth utilize One Stop Systems expansion hardware and composable infrastructure software solutions from Liqid and Dolphin to provide the dynamic reallocation of GPU and NVMe resources.

“We’re excited to be working with One Stop Systems’ expansion technology because it provides industry-leading GPU and NVMe hardware density, increasing resource availability across the data center,” said Jay Breakstone, CEO of Liqid. “Liqid’s technology platform allows users to manage, scale out, configure and even automate physical, bare-metal server systems. With the ability to treat GPUs as a disaggregated, shared resource for the first time, scaled via OSS expansion systems, composable solutions from Liqid deliver the infrastructure to meet today’s most taxing HPC challenges, such as peer-to-peer transfers and memory access for AI and machine learning.”

“Dolphin is pleased to partner with OSS to provide customers with the highest performance hardware and software,” said Hugo Kohmann, CEO of Dolphin. “The Dolphin eXpressWare SmartIO software offers a flexible way to enable PCIe IO devices such as NVMe, FPGAs and GPUs to be accessed within a PCIe Network. Devices can be borrowed over the PCIe network without any software overhead at the performance of PCI Express.”
Composable infrastructure is also available as a cloud solution. Data scientists can rent the latest technology composable infrastructure systems and software using operational expenditure budgets rather than capital equipment budgets. One Stop Systems is partnering with SkyScale to provide our composable infrastructure solutions in the cloud.

“Utilizing One Stop Systems hardware, SkyScale offers flexible composable infrastructure solutions in the Cloud,” said Tim Miller, President of SkyScale. “We already offer users cutting edge technology for GPU computing with unprecedented customization. By adding composable infrastructure solutions, we’re increasing the level of customization for our customers.”

Visitors to SC17 in Denver, Colorado can view these composable infrastructure demos in One Stop Systems’ booth #2049. Customers can order the hardware utilized in these demos from One Stop Systems’ highly-trained sales engineers.

About One Stop Systems

One Stop Systems designs and manufactures ultra-dense high performance computing (HPC) systems for deep learning, oil and gas exploration, financial trading, media and entertainment, defense and traditional HPC applications requiring the fastest and most efficient data processing. By utilizing the power of the latest GPU accelerators and NVMe flash cards, our systems stay on the cutting edge of the latest technologies. We have a reputation as innovators in hyperconverged and composable infrastructure solutions using the latest technology and design equipment to operate with the highest efficiency. Now OSS offers these exceptional systems to customers for lease or purchase. OSS continuously works to meet our customers’ greater needs.

Source: One Stop Systems

The post One Stop Systems Announces Composable Infrastructure Solutions at SC17 appeared first on HPCwire.


Subscribe to www.rmacc.org aggregator