BERKELEY LABORATORY RESEARCH COMPUTING
HPC Services offers comprehensive Linux cluster support, including pre-purchase consulting, procurement assistance, installation, and ongoing support for PI-owned clusters. Our HPC User Services consultants can help you to get your application running well to make best use of your new cluster. UC Berkeley PIs can also make use of our services through the new Berkeley Research Computing (BRC) program available through UCB Research IT.
Sep 10, 2014 - LabTech 2014 is here
Highlights of the day include our morning mini-classes, including 3 one hour sessions on getting the most out of Python in scientific computing, a 3 hour Arduino basics class, and, new this year, Intro and Advanced LabVIEW. The (free) Lunch and Keynote starts at , with an overview of what's new from IT this year. In the afternoon, you'll find over 30 sessions on topics ranging from Globus Online and HPC Use Cases to Video Conferencing to Excel. Please go here to register.
Aug 5, 2014 - Warewulf and NHC Talk at LCI Workshop
Senior HPC engineer Michael Jennings will be leading sessions on the LBNL-developed Warewulf cluster toolkit and Node Health Check at the 20th annual Linux Cluster Institute Workshop held from August 4-8, 2014 at the National Center for Supercomputing Applications (NCSA) in Urbana, Illinois.
May 22, 2014 - BERKELEY RESEARCH COMPUTING Launch Event
The UC Berkeley Research Computing (BRC) celebrates its launch on Thursday, May 22, 2014 at 3:00 p.m. in Sutardja Dai Hall. This exciting new program will include Condo/Institutional cluster (HPC) computing (in partnership with LBNL HPCS), cloud computing support, and virtual workstations. UC Berkeley researchers needing access to computation should look into this program. To attend the event, RSVP here
April 15, 2014 HPCS presenting at GlobusWorld 2014
LBL HPC consultant Krishna Muriki and HPCS systems engineer Karen Fernsler will be presenting at GlobusWorld 2014 on April 15-17, 2014 at Argonne National Laboratory.Their talk "Globus for Big Data and Science Gateways at LBL" will highlight some of our projects, including the Sloan Digital Sky Survey III and the X-Ray Diffraction and Microtomography Beamlines at the Advanced Light Source user facility, which benefit from Globus endpoints to construct Data Pipelines.
Feb 3, 2014 - NHC Talk at Stanford Exascale ConferenceLBL HPCS senior engineer Michael Jennings will be giving a talk on the "Node Health Check (NHC)" on Feb 4, 2014 at the Stanford Conference and Exascale Workshop 2014 sponsored by the HPC Advisory Council. NHC, developed by Jennings, provides the framework and implementation for a highly reliable, flexible, extensible node health check solution. It is now widely recommended by major HPC job scheduler vendors and is in use at many large HPC sites and research institutions.
Sept 6, 2013 - Monitoring with Ganglia book released
HPC Services cluster expert Bernard Li's new book "Monitoring with Ganglia - Tracking Dynamic Host and Application Metrics at Scale" has been recently published by O'Reilly media. His book shows how to use Ganglia to collect and visualize metrics from clusters, grids, and cloud infrastructures at any scale.
Mar 21, 2013 - Supporting Science with HPC
Mar 18, 2013 - GPU Accelerated Synchrotron Radiation Calculation
The climate models that scientists use to understand and project climate change are improving constantly; however the largest source of uncertainty in today’s climate models are clouds. As the source of rain and wind, clouds are important in modeling climate. Berkeley Lab scientist David Romps discusses his work to develop better cloud resolution models. More
Nov 13, 2012 - Node Health Check
Nov 12, 2012 - Warewulf wins Intel Cluster Ready "Explorer" award
Oct 24, 2012 - HPCS at 2012 Data Center Efficiency Summit
Sept 18, 2012 - Warewulf featured in HPC Admin Magazine
ALS physicists Changchun Sun and Hiroshi Nishimura along with HPCS staff Kai Song, Susan James, Krishna Muriki, Gary Jung, Bernard Li and Yong Qin recently explored the use of Amazon's VPC service to transparently extend the ALS compute cluster and software environment, into the public Cloud to provide on-demand compute resources for particle tracking and NGLS APEX development. Their work was presented during the poster session at the International Particle Accelerator Conference (IPAC12) in New Orleans this week.
HPC Services Manager Gary Jung talks about the issues institutions may encounter when developing new or enhancing existing infrastructure to support data intensive science at the Winter 2012 ESCC/Internet2 Joint Techs Conference in Baton Rouge, LA this week.
Researchers from Berkeley Lab and MIT have teamed up to develop a new tool,as part of the Materials Project, to speed up the development for new materials. The project incorporates the use of supercomputing resources including Lawrencium to characterize the properties of inorganic compounds. More
Oct 25, 2011 - Supercomputing As A Service
Big Data at the ALS
We build a Data Pipeline using a Fast 400MB/s CCD, a 43,002 core GPU cluster and a 15TB Data Transfer Node with Globus Online for PI David Shapiro to do the X-ray diffraction 3D image reconstruction at Beamline 9.0.1
Together with researchers in EETD, we evaluated the use of hot-water, direct-to-chip liquid cooling on a 800-core Intel IvyBridge cluster. Does it work well? Ask us.
Chemical Science Division One of our current projects is a Dell 720-core Intel IvyBridge processor cluster for DOE early career award winner Dan Haxton of the Atomic, Molecular, and Optical Theory Group
Material Science Division We recently completed a dedicated 1824-core Intel SandyBridge processor cluster, Catamount, for MSD researchers. The system will be used to investigate the development of advanced materials.
Joint Center for Artificial Photosynthesis (JCAP) We recently completed a dedicated 2944-core AMD Interlagos cluster for the researchers at JCAP, which is the nation's largest research program dedicated to the development of an artificial solar-fuel generation technology
Geologic Carbon Sequestration
A 64-node, 768-core, Condo expansion has been added to Lawrencium for PIs Quanlin Zhou and Curt Oldenburg in ESD. It will be used for modeling of geo carbon sequestration which is a major measure for climate change mitigation, and an important component of LBNL's Carbon Cycle 2.0 Initiative.