This resource is currently inactive.
This resource is currently disabled.
Resource Group
This Resource Group Belongs To
FacilityNational Energy Research Scientific Computing Center
This Resource Group is Supported By
Support CenterNERSC
Resource Description The NERSC Cray XT4 system, named Franklin, is a massively parallel processing (MPP) system with 9,660 compute nodes. Each node has quad processor cores, and the entire system has a total of 38,640 processor cores available for scientific applications. The system is named in honor of Benjamin Franklin. Each of Franklin\'s compute nodes consists of a 2.3 GHz single socket quad-core AMD Opteron processor (Budapest) with a theoretical peak performance of 9.2 GFlop/sec per core (4 flops/cycle if using SSE128 instructions). Each compute node has 8 GB of memory (2 GB of memory per core), and each service node (e.g. login node) has 8 GB of memory. The full system consists of 102 cabinets with 78 TBytes of aggregate memory. The memory speed is 800 MHz. The theoretical peak performance of Franklin is about 356 TFlop/sec. Each compute node is connected to a dedicated SeaStar2 router through Hypertransport with a 3D torus topology which ensures high performance, low-latency communication for MPI and SHMEM jobs. Franklin uses two different operating systems. Full-featured SuSE Linux is run on service nodes (10 login nodes and other I/O, network and system nodes). A light weight OS based on Linux, Compute Node Linux (CNL), is run on each compute node. CNL reduces system overhead, and is critical for the system to scale to very large concurrency. The overall operating system on Franklin is called Cray Linux Environment (CLE). The Parallel file system on Franklin is provided by Lustre with approximately 350 TB of user disk space.
Information URL
Resource FQDN
Compute Element
GridFtp Storage Element
VO Owners of This Resource
Submitter Contact A contact who has registered this resource
Shreyas Cholia
Resource Report Contact Gratia-accounting based resource-usage reports are sent out daily/weekly to these contacts
Shreyas Cholia
NERSC Grid Trouble (Shreyas Cholia)
Administrative Contact GOC tickets when assigned to a support center of a resource are also sent to the primary resource admin contact (usually the sysadmin of a resource) via the ticket's CC mechanism. Admin contacts are also contacted by GOC staff and others to deal with system-administrative problems
Shreyas Cholia
NERSC Grid Trouble (Shreyas Cholia)
Security Contact Security notifications sent out by the OSG security team are sent to primary and secondary security contacts for this resource
NERSC Security Group (Steve Chan)
Shreyas Cholia
WLCG Information
No WLCG Information
ActiveShows if this resource has been activated by GOC staff. Until it is activated, this resource is not available to OSG in general sense.False
DisableShows if this resource has been removed from the OIM database.True

Contact Ranks