Trinity (supercomputer)

Trinity
Trinity in August 2015
OperatorsNational Nuclear Security Administration
LocationLos Alamos National Laboratory
CostUS$174M[1]
PurposePrimarily utilized to perform milestone weapons calculations
Web sitelanl.gov/projects/trinity/

History

  • December 2013, The National Energy Research Scientific Computing Center (NERSC) and The Alliance for Computing at Extreme Scale (ACES) releases a joint RFP with technical requirements for Trinity.[2]
  • July 2014, Cray announces that they were awarded the $174 Million contract by the National Nuclear Security Administration to provide a next generation supercomputer to Los Alamos National Laboratory.[3]
  • June 2015, Haswell Partition installation begins.[4]
  • November 2015, Trinity appears on the Supercomputing Top500 list at #6.[5]
  • June 2016, Knights Landing Partition installation begins.[6]
  • November 2016, Trinity falls to #10 on the Top500 list.[7]
  • July 2017, The Haswell and KNL partitions are merged.[8]
  • November 2018, Trinity regains #6 spot on the Top500 list.[9]

Trinity technical specifications

Trinity High-Level Technical Specifications [10]
Operational Lifetime 2015 to 2020
Architecture Cray XC40
Memory Capacity 2.07 PiB
Peak Performance 41.5 PF/s
Number of Compute Nodes 19,420
Parallel File System Capacity 78 PB
Burst Buffer Capacity 3.7 PB
Footprint 4606 sq ft
power requirement 8.6 MW

Compute Tier

Trinity was built in 2 stages. The first stage incorporated the Intel Xeon Haswell processor while the second stage added a significant performance increase using the Intel Xeon Phi Knights Landing Processor. There are 301,952 Haswell and 678,912 Knights Landing processors in the combined system, yielding a total peak performance of over 40 PF/s (Petaflops per second)[11]

Storage Tiers

There are 5 primary storage tiers; Memory, Burst Buffer, Parallel File System, Campaign Storage, and Archive.[12]

Memory

2 PB of DDR4 DRAM provide physical memory for the machine. Each processor also has DRAM built on to the tile, providing additional memory capacity. The data in this tier is highly transient and is typically in residence for only a few seconds, being overwritten continuously. [13]

Burst Buffer

Cray supplies the three hundred XC40 Data Warp blades that each contain 2 Burst Buffer nodes and 4 SSD drives. There is a total of 3.78 PB of storage in this tier, capable of moving data at a rate of up to 2 TB/s. In this tier, data is typically resident for a few hours, with data being overwritten in approximately that same time frame. [14]

Parallel File System

Trinity uses a Sonexion based Lustre file system with a total capacity of 78 PB. Throughput on this tier is about 1.8 TB/s. It is used to stage data in preparation for HPC operations. Data residence in this tier is typically several weeks.

Campaign Storage

The MarFS Filesystem fits into the Campaign Storage tier and combines properties of POSIX and Object storage models. The capacity of this tier is growing at a rate of about 30 PB/year, with a current capacity of over 100 PB. In testing, LANL scientists were able to create 968 Billion files in a single directory at a rate of 835 million file creations per second. This storage is designed to be more robust than typical object storage, while sacrificing some of the end user functionality that you would get from a POSIX system. Performance of this tier is between 100-300 GB/s of throughput. Data residence in this tier is longer term, typically lasting several months.

Key Design goals

  • Transparency
  • Data protection
  • Recoverability
  • Ease of administration

MarFS is an open source filesystem and can be downloaded here: https://github.com/mar-file-system/marfs

Archive

The final layer of storage is the Archive. This is a HPSS tape file system that holds approximately 100 PB of data.

Infographic on Trinity's file storage system. Click to enlarge.
gollark: Turing Machineā„¢
gollark: BLC to PHP because it kind of rhymes!
gollark: A BLC to [random esolang] compiler?
gollark: Try compiling everything to Subtract and Branch if Negative!
gollark: LASM (Lambda (Calculus) Assembler)!

See also

References

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.