The FASRC Cannon cluster is a large-scale HPC (high performance computing) cluster supporting scientific modeling and simulation for thousands of Harvard researchers. Assembled with the support of the Faculty of Arts and Sciences, but since branching out to serve many Harvard units, it occupies more than 10,000 square feet with hundreds of racks spanning three data centers separated by 100 miles. The primary compute is housed in MGHPCC, our green (LEED Platinum) data center in Holyoke, MA. Other systems, including storage, login, virtual machines, and specialty compute, are housed in our Boston and Cambridge facilities.
CPU Compute: The new Cannon cluster is primarily comprised of 670 Lenovo SD650 NeXtScale servers, part of their new liquid-cooled Neptune line. Each chassis unit contains two nodes, each containing two Intel 8268 "Cascade Lake" processors and 192GB RAM per node. The nodes are interconnected by HDR 100 Gbps Infiniband (IB) in a single Fat Tree with a 200 Gbps IB core. The liquid cooling allows for efficient heat extraction while running higher clock speeds. Each Cannon node is several times faster than any previous Odyssey cluster node.
Additional to Cannon actual, nodes for various labs and projects also contribute to the total pool of resources available to the scheduler bringing the overall core count to ~ 100,000 cores.
GPU Compute: 16 Lenovo SR670 each with 32 CPU cores, 384GB memory, 4 x V100s (@ 5,120 CUDA cores each), connected via 100 Gbps InfiniBand. Additionally, hundreds of GPU-enabled server owned by faculty/labs are available via shared queues when idle.
Storage: FASRC now maintains over 40 PB of storage, and this keeps growing. Robust home directories are housed on enterprise-grade Isilon storage, while faster Lustre filesystems serve more performance-driven needs such as scratch and research shares. Our middle tier laboratory storage uses a mix of Lustre, Gluster and NFS filesystems. See our storage page for more details.
Interconnect: Odyssey has two underlying networks: A traditional TCP/IP network and low-latency InfiniBand networks that enable high-throughput messaging for inter-node parallel-computing and fast access to Lustre mounted storage. The IP network topology connects the three data centers together and presents them as a single contiguous environment to FASRC users.
Software: Our core operating system is CentOS. We maintain the configuration of the cluster and all related machines and services via Puppet. Cluster job scheduling is provided by SLURM (Simple Linux Utility for Resource Management) across several shared partitions, processing approximately 29,000,000 jobs per year. See our documentation on running jobs.
In addition to supporting cluster modules, we manage over a thousand different scientific software tools and programs. The FASRC user portal is a great place to find out more information on software modules, check on job status, and submit a help request. A license manager service is maintained for software requiring license checkout at run-time. Other services to the community include the distribution of individual use software packages and Citrix and Virtual OnDemand (VDI) instances that allow for remote display and execution of various software packages.
Hosted Machines: In addition to the HPC cluster, FASRC also manages hundreds of virtual machines for researchers and our own virtual infrastructure. These are provisioned conforming to three tiers of service, depending on the needs of the requester. Some use examples include data portals, database access points, and project development boxes. RC also provisions and manages core facility machines that are connected to the numerous instruments in labs for data collection and analysis.
Internal Infrastructure: FASRC also manages numerous machines which handle our internal infrastructure for authentication, account creation and management, documentation, help tickets, administration machines, out-of-band management, inventory, monitoring, statistics, and billing. FASRC supports over 5500 users in more than 500 labs.