(Available for LBNL researchers)
Overview
In recognition of the increasing importance of research computing across many scientific disciplines, LBNL has made a significant investment in developing the LBNL Condo Cluster program, as a way to grow and sustain midrange computing for Berkeley Lab. The Condo program is intended to provide Berkeley Lab researchers with state-of-the-art, professionally-administered computing systems and ancillary infrastructure, with the intent of improving competitiveness on grants, and achieving economies of scale with centralized computing systems and data center facilities.
The model for sustaining the Condo program is premised on faculty and principal investigators using equipment purchase funds from their grants or other available funds to purchase compute nodes (individual servers) which are then added to the Lab’s Lawrencium compute cluster. This allows PI-owned nodes to take advantage of the high speed Infiniband interconnect and high performance Lustre parallel filesystem storage associated with Lawrencium. Operating costs for managing and housing PI-owned compute nodes are waived in exchange for letting other users make use of any idle compute cycles on the PI-owned nodes. These PI owners have priority access to computing resources equivalent to those purchased with their funds, but now can access more nodes for their research if needed. This provides the PI with much greater flexibility as compared to owning a standalone cluster.
This program is intended for PIs that would otherwise purchase a small (4 nodes) to medium scale (72 nodes) standalone Linux cluster. Projects with larger compute needs or many users or groups should consider setting up a dedicated cluster so that they can better prioritize shared access between their users. Please request a condo account at myLRC with the relevant documents.
Program Details
Compute node equipment is purchased and maintained based on a 4-year lifecycle at which point the PI owning the nodes will be notified that the nodes will have to be upgraded during year 5. If the hardware is not upgraded by the end of 5 years, the PI may donate the equipment to Condo or take possession of the equipment (removal of the equipment from LC3 and transfer to another location is at the PI’s expense); nodes left in the cluster after five years may be removed and disposed of at the discretion of the HPCS program manager
All Lawrencium and condo users have a 30GB home directory on the Lab’s shared HPC infrastructure and are charged $25/mo. for account maintenance which includes backups of their home directory. Users or projects needing more space for persistent data can purchase storage shelves that can be hosted by the Lab’s HPC infrastructure. Storage shelves are purchased and maintained on a 5-yr lifecycle after which the PI must renew the storage purchase at the then-prevailing price or remove the data within 3 months.
Valid Project ID: To maintain your Lawrencium user account, you must have a valid PID. If your PID becomes invalid or you lose access to projects, for example, if you leave the last project or a PI removes you from it, your account will be temporarily blocked from logging into the cluster until a valid PID is supplied.
Once a PI has decided to participate, the PI or his designated person works with the HPC Services manager and operations team to procure the desired number of compute nodes and storage. Generally, it takes about three months from start to finish. In the interim, a test condo queue with a small allocation will be setup for the PI’s users in anticipation of the new equipment. Users may submit jobs to the general Lawrencium queues on the cluster, but use will incur the cpu usage fees of $0.01 per service unit. Jobs are subject to general queue limitations and guaranteed access to contributed cores is not provided until purchased nodes are provisioned.
Prospective condo owners should contact HPC Services Manager Wei Feinstein before purchasing any equipment to ensure compatibility.
Abdullah Cihan | Energy Geosciences |
Anubhav Jain | Energy Technologies Area |
Barry Freifeld | Earth and Environmental Sciences Area |
Curtis Oldenburg | Earth and Environmental Sciences Area |
Daniela Cassol | Joint Genome Institute |
David Limmer | Materials Sciences |
David Prendergast | Molecular Foundry |
David Romps | Earth and Environmental Sciences Area |
Erik Neuscamman | Chemical Sciences |
Gerbrand Ceder | Material Sciences Division |
Jeff Neaton | Molecular Foundry |
Jeff Porter | Nuclear Sciences Division |
Jens Birkholzer | Energy Geosciences Sciences Area |
Jin Qian | Chemical Sciences |
Joel Moore | Material Sciences Division |
Kjiersten Fagnan | Joint Genome Institute |
Kranthi Mandadapu | Chemical Sciences Division |
Kristin Persson | Energy Technologies Area |
Mark Asta | Material Sciences Division |
Martin Head-Gordon | Chemical Sciences Division |
Michael Zaletel | Materials Sciences |
Peter Lau | Earth and Environmental Sciences Area |
Phillip Geissler | Chemical Sciences Division |
Quanlin Zhou | Earth and Environmental Sciences Area |
Robert Lucchese | Chemical Sciences Division |
Samuel Blau | Energy Storage & Distributed R |
Sergi Molins | Energy Geosciences |
Teresa Head-Gordon | Chemical Sciences and Physical Bioscience Divisions |
William McCurdy | Chemical Sciences Division |