--- title: News and announcements linkTitle: Announcements type: docs weight: 1 aliases: - /news.html - /news --- Note: for the most current information on exceptions on HPCC's cluster please consult its [Alerts](/news/alerts) or [Twitter](https://twitter.com/UCR_HPCC) pages. ### Sep 2023 * Stata 17 has been superseded with Stata 18. Please update scripts relying on the stata/17 module to use the stata/18 module instead. ### Aug 2022 * In August 2022 we were awarded an MRI equipment grant ([#2215705](https://www.nsf.gov/awardsearch/showAward?AWD_ID=2215705&HistoricalAwards=false)) by NSF for the acquisition of a Big Data HPC Cluster in the total amount of $942,829. ### Feb/Mar 2022 * Rollout of Rocky and DUO: Feb 18 through Mar 17, 2022. For details see [here](https://hpcc.ucr.edu/changes/). ### Mar 2021 * Viet Pham joins HPCC as HPC systems administrator. - Welcome Viet! ### May/Jun 2020 * Melody Asghari joins HPCC as HPC systems administrator assistant. - Welcome Melody! ### Mar/Apr 2020 * For updates on HPCC's operation during the COVID-19 crisis please visit the [Alerts](/news/alerts#hpcc-operation-during-covid-19-crisis) page of this site. ### Sep 2019 * Travis Nasser joins HPCC as HPC systems administrator assistant. - Welcome Travis! * Major updates applied. - Check here for important [changes](/changes). ### Aug 2019 * Charles Forsyth left the position of HPC systems administrator and has moved on to new opportunities. - Good luck Chuck! ### Apr 2018 * Abraham Park joins HPCC as HPC systems administrator assistant. - Welcome Abe! ### Jun 2017 * Charles Forsyth joins HPCC as new full-time HPC systems administrator. - Welcome Chuck! ### Feb 2017 With funding provided by Michael Pazzani's office (RED) we were able to purchase and install major hardware upgrades. This included the following hardware resources: * Added 28 Intel nodes with a total of 896 CPU cores (or 1,792 logical CPU cores) and 512 GB of RAM per node * Added 8 NVIDIA K80 GPUs increasing total number of cuda cores in GPU queue to 59,904 * Redesign of Infiniband network to support new computer nodes and enhance future scalabilty of IB network to over 1000 nodes ### Dec 2016 * UCR approval of plans to form HPC Center ### Sept 2016 * Expansion of GPFS storage system: 2 disk enclosures for 120 8TB drives * Expansion of high-memory queue: 4 nodes * Install of new Intel batch queue: 12 nodes ### Mar 2016 * Expansion of batch queues: 14 nodes ### Apr 2015 * Deployment of new FDR IB network @ 56Gbs * Deployment of 28 AMD nodes (2,048 AMD cores), funded by NSF-MRI-2014 * Deployment of high-memory Intel nodes (each with 1TB RAM) * Deployment of GPU nodes (NVIDIA K80) * Deployment of big data GPFS disk storage system, funded by NIH-S10-2014 ### May 2014 Award of equipment grants from NSF and NIH * NIH-S10-2014 (1S10OD016290-01A1): $652,816 * NSF-MRI-2014 (MRI-1429826): $783,537