<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>FAS Research Computing Status - Incident history</title>
    <link>https://status.rc.fas.harvard.edu</link>
    <description>FAS Research Computing</description>
    <pubDate>Sun, 19 Apr 2026 11:55:47 +0000</pubDate>
    
<item>
  <title>login.rc.fas.harvard.edu is responding normally</title>
  <description>
    Type: Incident
    

    Affected Components: , Login Nodes - Holyoke, Login Nodes - Boston, , 
Login Nodes → 
login.rc.fas.harvard.edu →
    Apr 19, 11:55:47 GMT+0 - Investigating - login.rc.fas.harvard.edu is not responding normally. This incident was automatically created. Apr 19, 12:55:25 GMT+0 - Resolved - \\\[login.rc.fas.harvard.edu\\\](http://login.rc.fas.harvard.edu) is responding normally. This incident was automatically resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    
    <p><strong>Affected Components:</strong> , , , </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:55:47&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  login.rc.fas.harvard.edu is not responding normally. This incident was automatically created..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;12:55:25&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  \\\[login.rc.fas.harvard.edu\\\](http://login.rc.fas.harvard.edu) is responding normally. This incident was automatically resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Sun, 19 Apr 2026 11:55:47 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmo5pma7u0n7mjomagp7omaed</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmo5pma7u0n7mjomagp7omaed</guid>
</item>

<item>
  <title>Authentication outage</title>
  <description>
    Type: Incident
    Duration: 7 hours and 30 minutes

    Affected Components: Authentication
    Apr 19, 06:44:52 GMT+0 - Investigating - Authentication issues with openauth/radius. This incident was created by an automated monitoring service. Apr 19, 14:14:53 GMT+0 - Resolved - Openauth/radius is now operational. This update was created by an automated monitoring service. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 7 hours and 30 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;06:44:52&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  Authentication issues with openauth/radius. This incident was created by an automated monitoring service..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:14:53&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Openauth/radius is now operational. This update was created by an automated monitoring service..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Sun, 19 Apr 2026 06:44:52 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmo5eigd3025we79u8go90tos</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmo5eigd3025we79u8go90tos</guid>
</item>

<item>
  <title>MGHPCC Power Loss</title>
  <description>
    Type: Incident
    Duration: 8 hours and 54 minutes

    Affected Components: FASSE Compute Cluster (Holyoke), FASRC Two-Factor (OpenAuth), HolyLFS04 (Tier 0), GPU nodes (Holyoke), Infiniband - Holyoke/MGHPCC, SLURM Scheduler - FASSE, Holyoke Tier 2 NFS (new), Holylabs, Holyoke Firewall, Network - Holyoke/MGHPCC, HolyLFS06 (Tier 0), Holyoke/MGHPCC Data Center, Cannon Compute Cluster (Holyoke), seas_compute, Cannon Open OnDemand/VDI, FASSE login nodes, Kempner Cluster GPU, Kempner Cluster CPU, FASSE Open OnDemand/VDI, Virtual Infrastructure - Holyoke, Holystore01 (Tier 0), Login Nodes - Holyoke, NESE (NorthEast Storage Exchange), SLURM Scheduler - Cannon, Isilon Storage Holyoke (Tier 1), holECS, Holyoke Specialty Storage, Login Nodes - Boston, HolyLFS05 (Tier 0), Netscratch (Global Scratch), Tape - (Tier 3), Boston Compute Nodes
    Apr 19, 06:18:00 GMT+0 - Investigating - At 2:18am on April 19th MGHPCC (our Holyoke datacenter) lost cooling which caused the entire facility to shutdown. This caused the loss of all jobs that were running. Storage and data on that storage should be safe. The facility is working on restoring cooling and power. Unfortunately we do not have an ETA. Apr 19, 13:06:16 GMT+0 - Identified - Power was fully restored to MGHPCC at 7:39am on April 19th. FASRC staff has restored functionality to most systems except for FASSE Open OnDemand. All other services are up and operating normally. If you continue to see issues with any system that is marked operational please let us know. We will deal with any non urgent requests in normal working hours. Apr 19, 15:12:24 GMT+0 - Resolved - All services, including FASSE OOD, should be functional at this time. If you continue to see issues with any system that is marked operational please let us know by sending an email to [rchelp@rc.fas.harvard.edu](mailto:rchelp@rc.fas.harvard.edu)

We will deal with any non urgent requests in normal working hours.

This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 8 hours and 54 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;06:18:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  At 2:18am on April 19th MGHPCC (our Holyoke datacenter) lost cooling which caused the entire facility to shutdown. This caused the loss of all jobs that were running. Storage and data on that storage should be safe. The facility is working on restoring cooling and power. Unfortunately we do not have an ETA..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:06:16&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Power was fully restored to MGHPCC at 7:39am on April 19th. FASRC staff has restored functionality to most systems except for FASSE Open OnDemand. All other services are up and operating normally. If you continue to see issues with any system that is marked operational please let us know. We will deal with any non urgent requests in normal working hours..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:12:24&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  All services, including FASSE OOD, should be functional at this time. If you continue to see issues with any system that is marked operational please let us know by sending an email to [rchelp@rc.fas.harvard.edu](mailto:rchelp@rc.fas.harvard.edu)

We will deal with any non urgent requests in normal working hours.

This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Sun, 19 Apr 2026 06:18:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmo5mx8uq0wy7113vg0ey0q9t</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmo5mx8uq0wy7113vg0ey0q9t</guid>
</item>

<item>
  <title>Starfish down</title>
  <description>
    Type: Incident
    Duration: 1 day and 48 minutes

    Affected Components: Starfish
    Apr 13, 16:30:00 GMT+0 - Investigating - Starfish is currently unavailable, due to a network card issue. Updates to come.

We are currently investigating this incident. Apr 14, 14:17:21 GMT+0 - Identified - Staff will be at the datacenter today to check on the physical status of the server. Updates to come. 

We are continuing to work on a fix for this incident. Apr 14, 17:17:30 GMT+0 - Resolved - The network card has been replaced, and Starfish is back up. 

This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 day and 48 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 13&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:30:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  Starfish is currently unavailable, due to a network card issue. Updates to come.

We are currently investigating this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 14&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:17:21&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Staff will be at the datacenter today to check on the physical status of the server. Updates to come. 

We are continuing to work on a fix for this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 14&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:17:30&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  The network card has been replaced, and Starfish is back up. 

This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 13 Apr 2026 16:30:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnyotr4d003l6i6q561of3ku</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnyotr4d003l6i6q561of3ku</guid>
</item>

<item>
  <title>Coldfront is down.</title>
  <description>
    Type: Incident
    Duration: 43 minutes

    Affected Components: Coldfront
    Apr 8, 13:31:31 GMT+0 - Investigating - Coldfront logins are producing an error message. We are currently investigating this incident. Apr 8, 14:14:35 GMT+0 - Resolved - Coldfront is back up and accepting logins. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 43 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:31:31&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  Coldfront logins are producing an error message. We are currently investigating this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:14:35&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Coldfront is back up and accepting logins..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 8 Apr 2026 13:31:31 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnq3714o000vs7adgfjwe485</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnq3714o000vs7adgfjwe485</guid>
</item>

<item>
  <title>Starfish dashboard inaccessible</title>
  <description>
    Type: Incident
    Duration: 1 hour and 3 minutes

    Affected Components: Starfish
    Apr 6, 14:27:54 GMT+0 - Investigating - The Starfish dashboard is inaccessible. We are looking into the issue. Apr 6, 15:31:02 GMT+0 - Resolved - Starfish has resolved the issue and the dashboard is once again available. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 hour and 3 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:27:54&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The Starfish dashboard is inaccessible. We are looking into the issue..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:31:02&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Starfish has resolved the issue and the dashboard is once again available..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 6 Apr 2026 14:27:54 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnnabux30qmbq6fm9rbvrvx8</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnnabux30qmbq6fm9rbvrvx8</guid>
</item>

<item>
  <title>FASRC monthly maintenance April 6th 2026 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: FASRC Two-Factor (OpenAuth), , , Cannon Open OnDemand/VDI, FASSE login nodes, FASSE Open OnDemand/VDI, Login Nodes - Holyoke, Login Nodes - Boston, Netscratch (Global Scratch), , 
Login Nodes → 
VDI/OpenOnDemand → 
login.rc.fas.harvard.edu →
    Apr 6, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Apr 6, 17:00:00 GMT+0 - Completed - Maintenance has completed successfully Apr 6, 13:00:00 GMT+0 - Identified - FASRC monthly maintenance will take place on April 6th 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* [two-factor.rc.fas.harvard.edu](http://two-factor.rc.fas.harvard.edu) [OpenAuth](https://docs.rc.fas.harvard.edu/kb/openauth/) cut-over to new server  
   * Audience: New accounts or anyone requesting an OpenAuth token  
   * Impact: two-factor will be unavailable while moving to a new server
* RStudio Server (Open OnDemand)  
   * Audience: RStudio Server users on Cannon and FASSE  
   * Impact: We will be decommissioning some versions of RStudio Server so we can properly maintain all production versions. Versions to be decommissioned:  
         * R 4.1.3 (Bioconductor 3.14, RStudio 2022.02.0)  
         * R 4.1.0 (Bioconductor 3.13, RStudio 1.4.1717)  
         * R 4.0.3 (Bioconductor 3.12, Rstudio 1.3.1093)  
         * R 4.0.0 (Bioconductor 3.11, Rstudio 1.3.1093)  
   * If you use one of these versions, we recommend replacing it with the most recent version, R 4.4.2 (Bioconductor 3.20, RStudio 2024.12.0). You must reinstall previously installed libraries.
* Domain controller replacement  
   * Audience: Internal  
   * Impact: End users should not see any impact
* OOD/Open OnDemand reboots  
   * Audience: All OOD users, reboot of the head nodes  
   * Impact: Running sessions will _not_ be affected
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Netscratch 90-day retention cleanup  
   * Audience; All netscratch users  
   * Impact: Files older than 90 days will be removed per our [scratch policy](https://docs.rc.fas.harvard.edu/kb/policy-scratch/). Please note that this cleanup can happen at any time, not just during maintenance.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt; 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  FASRC monthly maintenance will take place on April 6th 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* [two-factor.rc.fas.harvard.edu](http://two-factor.rc.fas.harvard.edu) [OpenAuth](https://docs.rc.fas.harvard.edu/kb/openauth/) cut-over to new server  
   * Audience: New accounts or anyone requesting an OpenAuth token  
   * Impact: two-factor will be unavailable while moving to a new server
* RStudio Server (Open OnDemand)  
   * Audience: RStudio Server users on Cannon and FASSE  
   * Impact: We will be decommissioning some versions of RStudio Server so we can properly maintain all production versions. Versions to be decommissioned:  
         * R 4.1.3 (Bioconductor 3.14, RStudio 2022.02.0)  
         * R 4.1.0 (Bioconductor 3.13, RStudio 1.4.1717)  
         * R 4.0.3 (Bioconductor 3.12, Rstudio 1.3.1093)  
         * R 4.0.0 (Bioconductor 3.11, Rstudio 1.3.1093)  
   * If you use one of these versions, we recommend replacing it with the most recent version, R 4.4.2 (Bioconductor 3.20, RStudio 2024.12.0). You must reinstall previously installed libraries.
* Domain controller replacement  
   * Audience: Internal  
   * Impact: End users should not see any impact
* OOD/Open OnDemand reboots  
   * Audience: All OOD users, reboot of the head nodes  
   * Impact: Running sessions will _not_ be affected
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Netscratch 90-day retention cleanup  
   * Audience; All netscratch users  
   * Impact: Files older than 90 days will be removed per our [scratch policy](https://docs.rc.fas.harvard.edu/kb/policy-scratch/). Please note that this cleanup can happen at any time, not just during maintenance.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt;.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 6 Apr 2026 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmn6ac2960e2v140x1wt0rf80</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmn6ac2960e2v140x1wt0rf80</guid>
</item>

<item>
  <title>Scheduler is degraded</title>
  <description>
    Type: Incident
    Duration: 14 hours and 56 minutes

    Affected Components: , GPU nodes (Holyoke), , Cannon Compute Cluster (Holyoke), seas_compute, Kempner Cluster GPU, Kempner Cluster CPU, SLURM Scheduler - Cannon, Boston Compute Nodes, 
Cannon Cluster → 
Kempner Cluster →
    Apr 1, 12:11:15 GMT+0 - Resolved - This incident has been resolved. The scheduler is running normally. Mar 31, 21:15:24 GMT+0 - Investigating - The scheduler is in a degraded state due to [thrashing](https://en.wikipedia.org/wiki/Thrashing%5F%28computer%5Fscience%29)  
We are actively working to resolve this problem. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 14 hours and 56 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Apr &lt;var data-var=&#039;date&#039;&gt; 1&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;12:11:15&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved. The scheduler is running normally..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 31&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:15:24&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The scheduler is in a degraded state due to [thrashing](https://en.wikipedia.org/wiki/Thrashing%5F%28computer%5Fscience%29)  
We are actively working to resolve this problem..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Tue, 31 Mar 2026 21:15:24 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnf48s6u02wctz2d7zg0wjmn</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnf48s6u02wctz2d7zg0wjmn</guid>
</item>

<item>
  <title>two-factor.rc.fas.harvard.edu (openauth) error</title>
  <description>
    Type: Incident
    Duration: 1 hour and 12 minutes

    Affected Components: FASRC Two-Factor (OpenAuth)
    Mar 31, 15:32:20 GMT+0 - Investigating - We are currently investigating this incident. Requesting a new token or re-requesting your token from two-factor is not currently working.  Mar 31, 16:44:08 GMT+0 - Resolved - This incident has been resolved. two-factor.rc.fas.harvard.edu is working normally again. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 hour and 12 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 31&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:32:20&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  We are currently investigating this incident. Requesting a new token or re-requesting your token from two-factor is not currently working. .&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 31&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:44:08&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved. two-factor.rc.fas.harvard.edu is working normally again..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Tue, 31 Mar 2026 15:32:20 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnerzlwh0g8hr8lzi8oq5mr2</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnerzlwh0g8hr8lzi8oq5mr2</guid>
</item>

<item>
  <title>The web front end to two-factor.rc.fas.harvard.edu is currently not allowing logins, generating new tokens is currently unavailable</title>
  <description>
    Type: Incident
    Duration: 6 days and 30 minutes

    Affected Components: FASRC Two-Factor (OpenAuth)
    Mar 25, 14:30:00 GMT+0 - Investigating - We are currently investigating this incident. Mar 31, 15:00:25 GMT+0 - Resolved - This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 6 days and 30 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:30:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  We are currently investigating this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 31&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:00:25&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 25 Mar 2026 14:30:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmnes00kk0cz8e9ndj1i7o3xk</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmnes00kk0cz8e9ndj1i7o3xk</guid>
</item>

<item>
  <title>Network issues - Cluster degraded</title>
  <description>
    Type: Incident
    Duration: 5 days, 7 hours and 7 minutes

    Affected Components: , GPU nodes (Holyoke), Network - Holyoke/MGHPCC, Cannon Compute Cluster (Holyoke), seas_compute, SLURM Scheduler - Cannon, Isilon Storage Holyoke (Tier 1), Boston Compute Nodes, 
Cannon Cluster →
    Mar 25, 14:10:34 GMT+0 - Identified - Mounts to Holyoke Isilon (specifically /n/sw) are broken on numerous nodes across the cluster. We have a check rolling out to find these nodes so we can remediate them individually. Until remediated the cluster will be in a degraded state. Running jobs may randomly die or fail as they hit nodes that have stale mounts.

It will be risky to run jobs for the next hour and then, after that point, the cluster will have a large number of nodes closed waiting for them to drain so we can reboot them and fix the mounts. Mar 25, 13:34:01 GMT+0 - Investigating - A network issue affecting storage critical to the cluster is It&#039;s causing instability. The cluster is currently in a degraded state as a result. We are looking into the problem. Updates to follow.. Mar 25, 14:31:18 GMT+0 - Monitoring - Mounts to Holyoke Isilon (specifically /n/sw) are broken on numerous nodes across the cluster. We have a check rolling out to find these nodes so we can remediate them individually. Until remediated the cluster will be in a degraded state. Running jobs may randomly die or fail as they hit nodes that have stale mounts.

It will be risky to run jobs for the next hour and then, after that point, the cluster will have a large number of nodes closed waiting for them to drain so we can reboot them and fix the mounts.

At this time we are unaware of any holy-isilon problems other than the effect this had on cluster nodes/running jobs. We will update should we identify any data storage concerns. Mar 30, 20:41:25 GMT+0 - Resolved - This incident has been resolved by draining and rebooting any nodes with stuck mounts. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 5 days, 7 hours and 7 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:10:34&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Mounts to Holyoke Isilon (specifically /n/sw) are broken on numerous nodes across the cluster. We have a check rolling out to find these nodes so we can remediate them individually. Until remediated the cluster will be in a degraded state. Running jobs may randomly die or fail as they hit nodes that have stale mounts.

It will be risky to run jobs for the next hour and then, after that point, the cluster will have a large number of nodes closed waiting for them to drain so we can reboot them and fix the mounts..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:34:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  A network issue affecting storage critical to the cluster is It&#039;s causing instability. The cluster is currently in a degraded state as a result. We are looking into the problem. Updates to follow...&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:31:18&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Monitoring&lt;/strong&gt; -
  Mounts to Holyoke Isilon (specifically /n/sw) are broken on numerous nodes across the cluster. We have a check rolling out to find these nodes so we can remediate them individually. Until remediated the cluster will be in a degraded state. Running jobs may randomly die or fail as they hit nodes that have stale mounts.

It will be risky to run jobs for the next hour and then, after that point, the cluster will have a large number of nodes closed waiting for them to drain so we can reboot them and fix the mounts.

At this time we are unaware of any holy-isilon problems other than the effect this had on cluster nodes/running jobs. We will update should we identify any data storage concerns..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 30&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;20:41:25&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved by draining and rebooting any nodes with stuck mounts..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 25 Mar 2026 13:34:01 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmn634bsn0co6fzrz7gjvlmtv</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmn634bsn0co6fzrz7gjvlmtv</guid>
</item>

<item>
  <title>ColdFront is down.</title>
  <description>
    Type: Incident
    Duration: 1 hour and 54 minutes

    Affected Components: Coldfront
    Mar 19, 12:58:36 GMT+0 - Identified - ColdFront is down. We are working to bring it back up. The instance got replaced last night, but it had trouble configuring itself on the way up again. Mar 19, 14:52:54 GMT+0 - Resolved - Cold front is back up. Thank you for your patience. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 hour and 54 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;12:58:36&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  ColdFront is down. We are working to bring it back up. The instance got replaced last night, but it had trouble configuring itself on the way up again..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:52:54&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Cold front is back up. Thank you for your patience..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 19 Mar 2026 12:58:36 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmmxh7njd00qve87a87odxjyz</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmmxh7njd00qve87a87odxjyz</guid>
</item>

<item>
  <title>Key access issue to CSBN, HERS, FIINE, Portal Approve (p3approve)</title>
  <description>
    Type: Incident
    Duration: 5 days, 21 hours and 59 minutes

    Affected Components: portal.rc.fas.harvard.edu
    Mar 13, 20:35:07 GMT+0 - Investigating - We are currently investigating this incident. This only affects specific services. Users of CSBN, HERS, FIINE, Portal Approve (p3approve) may be affected. Email coming from these systems may also be delayed.  
No ETA Mar 19, 18:33:57 GMT+0 - Resolved - This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 5 days, 21 hours and 59 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 13&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;20:35:07&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  We are currently investigating this incident. This only affects specific services. Users of CSBN, HERS, FIINE, Portal Approve (p3approve) may be affected. Email coming from these systems may also be delayed.  
No ETA.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;18:33:57&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 13 Mar 2026 20:35:07 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmmpcvot800j93wkll4ba6gwu</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmmpcvot800j93wkll4ba6gwu</guid>
</item>

<item>
  <title>Starfish dashboard unavailable</title>
  <description>
    Type: Incident
    Duration: 1 hour and 16 minutes

    Affected Components: Starfish
    Mar 2, 16:22:42 GMT+0 - Investigating - The Starfish dashboard is not responding. We are currently investigating this issue with the vendor. Mar 2, 17:38:26 GMT+0 - Resolved - This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 hour and 16 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:22:42&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The Starfish dashboard is not responding. We are currently investigating this issue with the vendor..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:38:26&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 2 Mar 2026 16:22:42 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmm9e0oz109onj5z9yxz1dcjr</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmm9e0oz109onj5z9yxz1dcjr</guid>
</item>

<item>
  <title>FASRC monthly maintenance Monday March 2nd, 2026 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: Login Nodes - Holyoke, FASSE Compute Cluster (Holyoke), SLURM Scheduler - Cannon, , , GPU nodes (Holyoke), , SLURM Scheduler - FASSE, , Cannon Compute Cluster (Holyoke), , seas_compute, Cannon Open OnDemand/VDI, FASSE login nodes, Kempner Cluster CPU, Kempner Cluster GPU, FASSE Open OnDemand/VDI, Login Nodes - Boston, Netscratch (Global Scratch), , Boston Compute Nodes, 
Login Nodes → 
Cannon Cluster → 
VDI/OpenOnDemand → 
Kempner Cluster → 
FASSE Cluster → 
login.rc.fas.harvard.edu →
    Mar 2, 14:00:01 GMT+0 - Identified - Maintenance is now in progress Mar 2, 14:00:00 GMT+0 - Identified - Monthly maintenance will take place on Monday March 2nd, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm scheduler update  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* OOD node reboots  
   * Audience; All Open OnDemand users  
   * Impact: OOD nodes will reboot during the maintenance window
* Login node reboots  
   * Audience: All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Netscratch retention purge  
   * Audience: All users of Netscratch  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/) Mar 2, 18:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Monthly maintenance will take place on Monday March 2nd, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm scheduler update  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* OOD node reboots  
   * Audience; All Open OnDemand users  
   * Impact: OOD nodes will reboot during the maintenance window
* Login node reboots  
   * Audience: All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Netscratch retention purge  
   * Audience: All users of Netscratch  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/).&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;18:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 2 Mar 2026 14:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmls7rae90625fg9xy9b3jc5m</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmls7rae90625fg9xy9b3jc5m</guid>
</item>

<item>
  <title>Tape outage</title>
  <description>
    Type: Incident
    Duration: 4 days, 17 hours and 42 minutes

    Affected Components: NESE (NorthEast Storage Exchange), Tape - (Tier 3)
    Mar 4, 15:09:37 GMT+0 - Resolved - This incident has been resolved. Normal tape operations are restored. Feb 27, 21:27:09 GMT+0 - Investigating - NESE Tape service will be down or operating with degraded service (no store and recall) Friday from 12 Noon EST until as late as Monday, 2 March at 9 AM.  
  
SUMMARY OF ISSUE:  
  
NESE Tape service is currently not able to store or recall files to and from tape due to vendor firmware issues in the IBM TS4500 tape library. The issue is related to the library robotics and cartridge database and we do NOT expect any data loss from this issue.  
  
The issue is apparently due to an issue with the inventory database related to a recent firmware update. This database can be scrubbed and reconstructed by the library, which will scan the bar code labels on all the cartridges to rebuild the inventory. Association of files in Globus to tapes is handled separately from the tape library and is not affected by the firmware update. Mar 2, 14:03:01 GMT+0 - Identified - NESE Tape Service is still working with IBM technical support at restoring the inventory. The expected downtime is extended until Tuesday March 3rd, 9am.  
Apologies for the inconvenvenience. Mar 3, 14:04:46 GMT+0 - Monitoring - The tape library outage is further extended to Wednesday March 4th at 9am awaiting a hardware replacement part due today. Data can still be uploaded to lab collections via Globus, but be mindful of the 10 TB buffer file limit. The outage affects storage and recall from tape. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 4 days, 17 hours and 42 minutes</p>
    <p><strong>Affected Components:</strong> , </p>
    &lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 4&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:09:37&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved. Normal tape operations are restored..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 27&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:27:09&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  NESE Tape service will be down or operating with degraded service (no store and recall) Friday from 12 Noon EST until as late as Monday, 2 March at 9 AM.  
  
SUMMARY OF ISSUE:  
  
NESE Tape service is currently not able to store or recall files to and from tape due to vendor firmware issues in the IBM TS4500 tape library. The issue is related to the library robotics and cartridge database and we do NOT expect any data loss from this issue.  
  
The issue is apparently due to an issue with the inventory database related to a recent firmware update. This database can be scrubbed and reconstructed by the library, which will scan the bar code labels on all the cartridges to rebuild the inventory. Association of files in Globus to tapes is handled separately from the tape library and is not affected by the firmware update..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:03:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  NESE Tape Service is still working with IBM technical support at restoring the inventory. The expected downtime is extended until Tuesday March 3rd, 9am.  
Apologies for the inconvenvenience..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Mar &lt;var data-var=&#039;date&#039;&gt; 3&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:04:46&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Monitoring&lt;/strong&gt; -
  The tape library outage is further extended to Wednesday March 4th at 9am awaiting a hardware replacement part due today. Data can still be uploaded to lab collections via Globus, but be mindful of the 10 TB buffer file limit. The outage affects storage and recall from tape..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 27 Feb 2026 21:27:09 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmm5ekmnr0020fzjvxkjv1iox</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmm5ekmnr0020fzjvxkjv1iox</guid>
</item>

<item>
  <title>Starfish dashboard is unavailable</title>
  <description>
    Type: Incident
    Duration: 1 day, 7 hours and 50 minutes

    Affected Components: Starfish
    Feb 27, 22:04:05 GMT+0 - Resolved - This incident has been resolved. The Starfish dashboard is available. Feb 26, 14:13:35 GMT+0 - Investigating - The starfish dashboard is unavailable. We are currently investigating this issue with Starfish.. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 day, 7 hours and 50 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 27&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;22:04:05&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved. The Starfish dashboard is available..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 26&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:13:35&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The starfish dashboard is unavailable. We are currently investigating this issue with Starfish...&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 26 Feb 2026 14:13:35 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmm3jn9o500pvn3vlsrovaqif</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmm3jn9o500pvn3vlsrovaqif</guid>
</item>

<item>
  <title>Starfish maintenance Feb 25, 2026 all day</title>
  <description>
    Type: Maintenance
    Duration: 1 day

    Affected Components: Starfish
    Feb 25, 14:00:00 GMT+0 - Identified - Starfish will be unavailable starting Wednesday, February 25th at 9AM until Thursday, February 26th at 9AM, for routine maintenance. The online dashboard will be inaccessible during this time. Feb 26, 14:00:00 GMT+0 - Completed - Maintenance has completed successfully Feb 25, 14:00:01 GMT+0 - Identified - Maintenance is now in progress 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 1 day</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Starfish will be unavailable starting Wednesday, February 25th at 9AM until Thursday, February 26th at 9AM, for routine maintenance. The online dashboard will be inaccessible during this time..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 26&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 25&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 25 Feb 2026 14:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmlvay24p0nlme0oeqgve2zzk</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmlvay24p0nlme0oeqgve2zzk</guid>
</item>

<item>
  <title>Authentication outage</title>
  <description>
    Type: Incident
    Duration: 5 minutes

    Affected Components: Authentication
    Feb 24, 15:39:56 GMT+0 - Investigating - Authentication issues with openauth/radius. This incident was created by an automated monitoring service. Feb 24, 15:44:57 GMT+0 - Resolved - Openauth/radius is now operational. This update was created by an automated monitoring service. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 5 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 24&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:39:56&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  Authentication issues with openauth/radius. This incident was created by an automated monitoring service..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 24&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:44:57&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Openauth/radius is now operational. This update was created by an automated monitoring service..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Tue, 24 Feb 2026 15:39:56 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmm0ruk3v0117ca6jfq7m0o6l</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmm0ruk3v0117ca6jfq7m0o6l</guid>
</item>

<item>
  <title>NESE tape maintenance Feb 19th 2026</title>
  <description>
    Type: Maintenance
    Duration: 9 hours

    Affected Components: NESE (NorthEast Storage Exchange)
    Feb 19, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Feb 19, 13:00:00 GMT+0 - Identified - From our partners at NESE. Details follow:

We are installing four new tape frames, which will bring the tape system raw storage capacity to 253 petabytes.

**Service Affected:** NESE Tape Service

**Maintenance Window:** 8:00 AM - 5:00 PM (EST)

* The tape service will be unavailable.
* All upgrade activities are expected to be completed on the same day.

NOTES:

* Monitor the MGHPCC Slack #nese channel for status updates and announcements
* Monitor &lt;https://nese.instatus.com/&gt; for real-time updates on progress

Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements Feb 19, 22:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 9 hours</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  From our partners at NESE. Details follow:

We are installing four new tape frames, which will bring the tape system raw storage capacity to 253 petabytes.

**Service Affected:** NESE Tape Service

**Maintenance Window:** 8:00 AM - 5:00 PM (EST)

* The tape service will be unavailable.
* All upgrade activities are expected to be completed on the same day.

NOTES:

* Monitor the MGHPCC Slack #nese channel for status updates and announcements
* Monitor &lt;https://nese.instatus.com/&gt; for real-time updates on progress

Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 19&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;22:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 19 Feb 2026 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmkx2dbd201zt5svdsbm0pm92</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmkx2dbd201zt5svdsbm0pm92</guid>
</item>

<item>
  <title>OOD inaccessible</title>
  <description>
    Type: Incident
    Duration: 3 hours and 30 minutes

    Affected Components: , Cannon Open OnDemand/VDI, FASSE Open OnDemand/VDI, 
VDI/OpenOnDemand →
    Feb 11, 19:45:06 GMT+0 - Resolved - This incident has been resolved and OOD is working normally. Feb 11, 16:15:00 GMT+0 - Investigating - OpenOnDemand for both Cannon and FASSE may be inaccessible for some users. Errors may include: 

&quot;Error -- can&#039;t find user for &lt;username&gt;&quot;

&quot;502 proxy errors&quot;

For users that are able to access OOD, performance may be degraded or sessions may get stuck. 

We are currently investigating the root causes of this incident. Updates to follow  
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 3 hours and 30 minutes</p>
    <p><strong>Affected Components:</strong> , , </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 11&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;19:45:06&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved and OOD is working normally..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 11&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:15:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  OpenOnDemand for both Cannon and FASSE may be inaccessible for some users. Errors may include: 

&quot;Error -- can&#039;t find user for &lt;username&gt;&quot;

&quot;502 proxy errors&quot;

For users that are able to access OOD, performance may be degraded or sessions may get stuck. 

We are currently investigating the root causes of this incident. Updates to follow .&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 11 Feb 2026 16:15:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmlidbff700e2v2z4sn8zkflz</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmlidbff700e2v2z4sn8zkflz</guid>
</item>

<item>
  <title>Security updates needed for www.rc.fas.harvard.edu and docs.rc.fas.harvard.edu</title>
  <description>
    Type: Maintenance
    Duration: 8 minutes

    Affected Components: docs.rc.fas.harvard.edu, www.rc.fas.harvard.edu
    Feb 9, 21:10:00 GMT+0 - Identified - Security updates will require a brief interruption for our primary websites [www.rc.fas.harvard.edu](http://www.rc.fas.harvard.edu) and [docs.rc.fas.harvard.edu](http://docs.rc.fas.harvard.edu)

We will endeavour to keep this update as short as possible. Each site may be unavailable for a few minutes. Feb 9, 21:10:01 GMT+0 - Identified - Maintenance is now in progress Feb 9, 21:17:51 GMT+0 - Completed - Maintenance has completed successfully. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 8 minutes</p>
    <p><strong>Affected Components:</strong> , </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:10:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Security updates will require a brief interruption for our primary websites [www.rc.fas.harvard.edu](http://www.rc.fas.harvard.edu) and [docs.rc.fas.harvard.edu](http://docs.rc.fas.harvard.edu)

We will endeavour to keep this update as short as possible. Each site may be unavailable for a few minutes..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:10:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:17:51&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 9 Feb 2026 21:10:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmlfnw5qt0xla10jy3plyxdx0</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmlfnw5qt0xla10jy3plyxdx0</guid>
</item>

<item>
  <title>License server issue</title>
  <description>
    Type: Incident
    Duration: 18 minutes

    Affected Components: Software &amp; Modules, , FIINE billing portal, License Servers, Starfish, FASRC Downloads Site, Citrix, Grafana Cloud (FASRC), docs.rc.fas.harvard.edu, portal.rc.fas.harvard.edu, Spinal, Bauer MiniLIMS, www.rc.fas.harvard.edu, FASRC Offsite Hosting, FASRC Ticket System (ServiceNow), Coldfront, 
Websites &amp; Tools →
    Feb 9, 18:54:59 GMT+0 - Investigating - New sessions of Matlab are hanging. 

We are currently investigating this incident. Feb 9, 19:04:52 GMT+0 - Identified - The affected softwares include: 

Matlab

Mathematica

Gurobi

We are continuing to work on a fix for this incident. Feb 9, 19:13:20 GMT+0 - Resolved - The license server is back up, and all software should be performing as expected. 

This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 18 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;18:54:59&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  New sessions of Matlab are hanging. 

We are currently investigating this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;19:04:52&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  The affected softwares include: 

Matlab

Mathematica

Gurobi

We are continuing to work on a fix for this incident..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;19:13:20&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  The license server is back up, and all software should be performing as expected. 

This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 9 Feb 2026 18:54:59 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmlfj7meu0zfk5p7vmpixtiwd</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmlfj7meu0zfk5p7vmpixtiwd</guid>
</item>

<item>
  <title>NESE tape maintenance Feb 9th 2026</title>
  <description>
    Type: Maintenance
    Duration: 9 hours

    Affected Components: NESE (NorthEast Storage Exchange)
    Feb 9, 13:00:00 GMT+0 - Identified - From our partners at NESE. Details follow:

In the process of the tape front-end file caching system upgrade, we will be installing a new IBM Storage Scale System 6000\. We will provide an additional update for when the software integration and data transfer from the current IBM Elastic Storage System 5000 will be performed.

**Service Affected:** NESE Tape Service

**Maintenance Window: No Downtime expected**

NOTES:

* Monitor the MGHPCC Slack #nese channel for status updates and announcements
* Monitor &lt;https://nese.instatus.com/&gt; for real-time updates on progress
* Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements Feb 9, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Feb 9, 22:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 9 hours</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  From our partners at NESE. Details follow:

In the process of the tape front-end file caching system upgrade, we will be installing a new IBM Storage Scale System 6000\. We will provide an additional update for when the software integration and data transfer from the current IBM Elastic Storage System 5000 will be performed.

**Service Affected:** NESE Tape Service

**Maintenance Window: No Downtime expected**

NOTES:

* Monitor the MGHPCC Slack #nese channel for status updates and announcements
* Monitor &lt;https://nese.instatus.com/&gt; for real-time updates on progress
* Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;22:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 9 Feb 2026 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmkx2d5pn01lsytjor57r552r</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmkx2d5pn01lsytjor57r552r</guid>
</item>

<item>
  <title>Grafana Cloud (FASRC) is down</title>
  <description>
    Type: Incident
    

    Affected Components: Grafana Cloud (FASRC)
    Feb 6, 20:44:11 GMT+0 - Investigating - Grafana Cloud (FASRC) is down at the moment. This incident was automatically created by Instatus monitoring. Feb 6, 20:48:52 GMT+0 - Resolved - Grafana Cloud (FASRC) is back up. This incident was automatically resolved by Instatus monitoring. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;20:44:11&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  Grafana Cloud (FASRC) is down at the moment. This incident was automatically created by Instatus monitoring..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;20:48:52&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Grafana Cloud (FASRC) is back up. This incident was automatically resolved by Instatus monitoring..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 6 Feb 2026 20:44:10 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmlbcsgmd000u3t5jsy7daltl</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmlbcsgmd000u3t5jsy7daltl</guid>
</item>

<item>
  <title>FASRC monthly maintenance Monday February 2nd, 2026 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: Cannon Open OnDemand/VDI, Login Nodes - Holyoke, , , FASSE Compute Cluster (Holyoke), GPU nodes (Holyoke), , SLURM Scheduler - FASSE, , SLURM Scheduler - Cannon, Cannon Compute Cluster (Holyoke), , FASSE Open OnDemand/VDI, seas_compute, FASSE login nodes, Kempner Cluster CPU, Kempner Cluster GPU, Login Nodes - Boston, Netscratch (Global Scratch), , Boston Compute Nodes, 
Login Nodes → 
Cannon Cluster → 
VDI/OpenOnDemand → 
Kempner Cluster → 
FASSE Cluster → 
login.rc.fas.harvard.edu →
    Feb 2, 14:00:00 GMT+0 - Identified - Monthly maintenance will take place on Monday February 2nd, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* MaxTime change  
   * Audience: Cluster users  
   * Impact: In order to improve scheduling efficiency and stability, we will be setting a maximum run time on all partitions that have MaxTime set to UNLIMITED to a MaxTime of 3 days. The unrestricted partition will be set to 365 days. Partitions that already have MaxTime set will retain their current setting. Partition owners wishing to set a different MaxTime for their partition should contact FASRC. Note that we do no guarantee uptime and so users should utilize checkpointing to save state in case of node failure.
* Slurm upgrade to 25.11.2  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* OOD node reboots  
   * Audience; All Open OnDemand users  
   * Impact: OOD nodes will reboot during the maintenance window
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/) Feb 2, 14:00:01 GMT+0 - Identified - Maintenance is now in progress Feb 2, 18:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Monthly maintenance will take place on Monday February 2nd, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* MaxTime change  
   * Audience: Cluster users  
   * Impact: In order to improve scheduling efficiency and stability, we will be setting a maximum run time on all partitions that have MaxTime set to UNLIMITED to a MaxTime of 3 days. The unrestricted partition will be set to 365 days. Partitions that already have MaxTime set will retain their current setting. Partition owners wishing to set a different MaxTime for their partition should contact FASRC. Note that we do no guarantee uptime and so users should utilize checkpointing to save state in case of node failure.
* Slurm upgrade to 25.11.2  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* OOD node reboots  
   * Audience; All Open OnDemand users  
   * Impact: OOD nodes will reboot during the maintenance window
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/).&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Feb &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;18:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 2 Feb 2026 14:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmkvgvpg4095lzxnkkb2spait</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmkvgvpg4095lzxnkkb2spait</guid>
</item>

<item>
  <title>Coldfront down</title>
  <description>
    Type: Incident
    Duration: 59 minutes

    Affected Components: Coldfront
    Jan 22, 16:04:46 GMT+0 - Resolved - Coldfront is operational. Thank you for your patience. Jan 22, 15:06:02 GMT+0 - Investigating - We are currently investigating an issues with Coldfront. No ETA. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 59 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 22&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:04:46&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  Coldfront is operational. Thank you for your patience..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 22&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:06:02&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  We are currently investigating an issues with Coldfront. No ETA..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 22 Jan 2026 15:06:02 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmkpl3ux008buvdviozj30q33</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmkpl3ux008buvdviozj30q33</guid>
</item>

<item>
  <title>holystore01 down</title>
  <description>
    Type: Incident
    Duration: 1 hour and 1 minute

    Affected Components: Holystore01 (Tier 0)
    Jan 21, 16:28:00 GMT+0 - Investigating - The filesystem holystore01 is experiencing a network failure and in a bad state. 

Some files on holystore01 may not be accessible while this is ongoing. We are working to restore access, and apologize for the inconvenience.  Jan 21, 17:29:01 GMT+0 - Resolved - holystore01 is back up and usable. 

This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 1 hour and 1 minute</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 21&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:28:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The filesystem holystore01 is experiencing a network failure and in a bad state. 

Some files on holystore01 may not be accessible while this is ongoing. We are working to restore access, and apologize for the inconvenience. .&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 21&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:29:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  holystore01 is back up and usable. 

This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 21 Jan 2026 16:28:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmko8lfbe001zmrieugg7o6iy</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmko8lfbe001zmrieugg7o6iy</guid>
</item>

<item>
  <title>Grafana Cloud is down</title>
  <description>
    Type: Incident
    

    
    Jan 15, 15:34:57 GMT+0 - Investigating -  is down at the moment. This incident was automatically created by Instatus monitoring. Jan 15, 15:46:56 GMT+0 - Resolved -  is back up. This incident was automatically resolved by Instatus monitoring. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    
    
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 15&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:34:57&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
   is down at the moment. This incident was automatically created by Instatus monitoring..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 15&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:46:56&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
   is back up. This incident was automatically resolved by Instatus monitoring..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 15 Jan 2026 15:34:57 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmkfm22ra00gx7uwlws14tkm0</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmkfm22ra00gx7uwlws14tkm0</guid>
</item>

<item>
  <title>holystore01 is wedging. We are rebooting.</title>
  <description>
    Type: Incident
    Duration: 14 minutes

    Affected Components: Holystore01 (Tier 0)
    Jan 13, 14:57:56 GMT+0 - Identified - holystore01 is wedging. We are rebooting. Jan 13, 15:12:13 GMT+0 - Resolved - This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 14 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 13&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:57:56&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  holystore01 is wedging. We are rebooting..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 13&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:12:13&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Tue, 13 Jan 2026 14:57:56 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmkcpurir09emrh2dgobpkwsc</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmkcpurir09emrh2dgobpkwsc</guid>
</item>

<item>
  <title>FASRC monthly maintenance Monday January 12th, 2026 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: , , FASSE Compute Cluster (Holyoke), SLURM Scheduler - Cannon, GPU nodes (Holyoke), , SLURM Scheduler - FASSE, , Cannon Compute Cluster (Holyoke), , Cannon Open OnDemand/VDI, FASSE Open OnDemand/VDI, seas_compute, Login Nodes - Holyoke, FASSE login nodes, Kempner Cluster CPU, Kempner Cluster GPU, Login Nodes - Boston, , Boston Compute Nodes, 
Login Nodes → 
Cannon Cluster → 
VDI/OpenOnDemand → 
Kempner Cluster → 
FASSE Cluster → 
login.rc.fas.harvard.edu →
    Jan 12, 14:00:00 GMT+0 - Identified - Monthly maintenance will take place on January 12th, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Changes to SEAS partitions, please see tasks below.
* Changes to job age priority weighting, please see tasks below.
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?:**YES**

* Slurm upgrade to 25.11.1  
   * Audience: All cluster users (Cannon and FASSE)  
   * Impact: Jobs will be paused during maintenance
* In conjunction with SEAS we will modify seas\_gpu and seas\_compute time limits  
   * Audience: SEAS users  
   * Impact:  
   seas\_gpu: will be set to 2 days maximum  
   seas\_compute: will be set to 3 days maximum  
   Existing pending jobs longer than these limits will be set to 2 day and 3 day run times depending on partition.
* Job Age Priority Weight Change  
   * Audience: Cluster users  
   * Impact: We will be adjusting the weight applied to the priority earned by jobs by virtue of their age. Currently job priority is made up of two factors, Fairshare and Job Age. The Job Age factor is currently set such that jobs gain priority over 3 days with a maximum priority equivalent to jobs with Fairshare of 0.5\. This keeps low fairshare jobs from languishing at the bottom of the queue. With the current settings though, users with low fairshare can gain significant advantage over users with higher relative fairshare. To remedy this we will be adjusting the Job Age weight to cap out at an equivalent Fairshare of 0.1\. This will still allow jobs with 0 fairshare to gain priority and thus not languish while letting fairshare govern a wider range of higher priority jobs.
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Open OnDemand (OOD) node reboots  
   * Audienc:; All OOD users  
   * Impact: OOD nodes will reboot during the maintenance window
* Netscratch retention will run  
   * Audience: All cluster netscratch users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/) Jan 12, 14:00:01 GMT+0 - Identified - Maintenance is now in progress Jan 12, 18:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 12&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Monthly maintenance will take place on January 12th, 2026\. Our maintenance tasks should be completed between 9am-1pm.

**NOTICES:**

* Changes to SEAS partitions, please see tasks below.
* Changes to job age priority weighting, please see tasks below.
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?:**YES**

* Slurm upgrade to 25.11.1  
   * Audience: All cluster users (Cannon and FASSE)  
   * Impact: Jobs will be paused during maintenance
* In conjunction with SEAS we will modify seas\_gpu and seas\_compute time limits  
   * Audience: SEAS users  
   * Impact:  
   seas\_gpu: will be set to 2 days maximum  
   seas\_compute: will be set to 3 days maximum  
   Existing pending jobs longer than these limits will be set to 2 day and 3 day run times depending on partition.
* Job Age Priority Weight Change  
   * Audience: Cluster users  
   * Impact: We will be adjusting the weight applied to the priority earned by jobs by virtue of their age. Currently job priority is made up of two factors, Fairshare and Job Age. The Job Age factor is currently set such that jobs gain priority over 3 days with a maximum priority equivalent to jobs with Fairshare of 0.5\. This keeps low fairshare jobs from languishing at the bottom of the queue. With the current settings though, users with low fairshare can gain significant advantage over users with higher relative fairshare. To remedy this we will be adjusting the Job Age weight to cap out at an equivalent Fairshare of 0.1\. This will still allow jobs with 0 fairshare to gain priority and thus not languish while letting fairshare govern a wider range of higher priority jobs.
* Login node reboots  
   * Audience; All login node users  
   * Impact: Login nodes will reboot during the maintenance window
* Open OnDemand (OOD) node reboots  
   * Audienc:; All OOD users  
   * Impact: OOD nodes will reboot during the maintenance window
* Netscratch retention will run  
   * Audience: All cluster netscratch users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
[https://www.rc.fas.harvard.edu/](https://www.rc.fas.harvard.edu/upcoming-training/).&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 12&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 12&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;18:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 12 Jan 2026 14:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmk1ivije002e49jtjj5n83yl</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmk1ivije002e49jtjj5n83yl</guid>
</item>

<item>
  <title>fasselogin01 reboot</title>
  <description>
    Type: Incident
    Duration: 26 minutes

    Affected Components: FASSE login nodes
    Jan 9, 16:32:56 GMT+0 - Identified - fasselogin01 will be rebooted at 11:45 to fix some mounts Jan 9, 16:58:42 GMT+0 - Resolved - fasselogin01 is back with the correct mounts. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 26 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:32:56&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  fasselogin01 will be rebooted at 11:45 to fix some mounts.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:58:42&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  fasselogin01 is back with the correct mounts..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 9 Jan 2026 16:32:56 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmk73hin4089gm6zlauy9ilxl</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmk73hin4089gm6zlauy9ilxl</guid>
</item>

<item>
  <title>Slurm down</title>
  <description>
    Type: Incident
    Duration: 2 hours and 7 minutes

    Affected Components: NESE (NorthEast Storage Exchange), Holyoke Tier 2 NFS (new), Holyoke-Boston fiber link (long path), Globus Data Transfer, GPU nodes (Holyoke), FASSE Compute Cluster (Holyoke), Isilon Storage Holyoke (Tier 1), Starfish, Boston Tier 2 NFS (new), Infiniband - Holyoke/MGHPCC, HolyLFS04 (Tier 0), SLURM Scheduler - FASSE, Web Proxies, Network - Holyoke/MGHPCC, Virtual Infrastructure - Holyoke, Holyoke-Boston fiber link (short path), Infiniband - Boston, HolyLFS06 (Tier 0), bosECS, Boston Specialty Storage, Authentication, Boston Data Center, Samba Cluster, Software &amp; Modules, Holyoke/MGHPCC Data Center, Cannon Compute Cluster (Holyoke), Network - Boston, CEPH Storage Boston (Tier 2), Cannon Open OnDemand/VDI, Cambridge firewall and other redundancy, SLURM Scheduler - Cannon, Holylabs, FASSE Open OnDemand/VDI, FIINE billing portal, License Servers, seas_compute, Login Nodes - Holyoke, FASSE login nodes, Kempner Cluster CPU, Kempner Cluster GPU, holECS, Holyoke Specialty Storage, FASRC Two-Factor (OpenAuth), FASRC Downloads Site, Citrix, Login Nodes - Boston, HolyLFS05 (Tier 0), Virtual Infrastructure - Boston, Network - Cambridge, FASRC VPN (Cambridge) , FASRC VPN (Boston), Holystore01 (Tier 0), Grafana Cloud (FASRC), BosLFS02 (Tier 0), Isilon Storage Boston (Tier 1), Home Directory Storage - Boston, docs.rc.fas.harvard.edu, portal.rc.fas.harvard.edu, Spinal, Bauer MiniLIMS, www.rc.fas.harvard.edu, Netscratch (Global Scratch), Tape - (Tier 3), FASRC Offsite Hosting, FASRC Ticket System (ServiceNow), Boston Compute Nodes, Harvard DNS System, Coldfront, Holyoke Firewall
    Jan 9, 00:11:25 GMT+0 - Investigating - The Slurm scheduler is currently down and no new jobs are able to be scheduled. 

We are currently investigating this incident and will provide updates.  Jan 9, 02:18:39 GMT+0 - Resolved - The rogue job has been found and removed. The scheduler is running normally again and all partitions are open. Jan 9, 02:20:18 GMT+0 - Resolved - This incident has been resolved. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Incident</p>
    <p><strong>Duration:</strong> 2 hours and 7 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;00:11:25&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Investigating&lt;/strong&gt; -
  The Slurm scheduler is currently down and no new jobs are able to be scheduled. 

We are currently investigating this incident and will provide updates. .&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;02:18:39&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  The rogue job has been found and removed. The scheduler is running normally again and all partitions are open..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jan &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;02:20:18&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Resolved&lt;/strong&gt; -
  This incident has been resolved..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 9 Jan 2026 00:11:25 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/incident/cmk64fa0902qk10oyoj43y58n</link>
  <guid>https://status.rc.fas.harvard.edu/incident/cmk64fa0902qk10oyoj43y58n</guid>
</item>

<item>
  <title>Monthly Maintenance and MGHPCC Power Work - Dec. 8, 2025 6am-6pm</title>
  <description>
    Type: Maintenance
    Duration: 12 hours

    Affected Components: Login Nodes - Holyoke, , GPU nodes (Holyoke), FASSE Compute Cluster (Holyoke), Isilon Storage Holyoke (Tier 1), , SLURM Scheduler - FASSE, Virtual Infrastructure - Holyoke, , Cannon Compute Cluster (Holyoke), , Cannon Open OnDemand/VDI, SLURM Scheduler - Cannon, FASSE Open OnDemand/VDI, License Servers, seas_compute, , FASSE login nodes, Kempner Cluster CPU, Kempner Cluster GPU, Login Nodes - Boston, Virtual Infrastructure - Boston, Isilon Storage Boston (Tier 1), , Boston Compute Nodes, 
Login Nodes → 
VDI/OpenOnDemand → 
Kempner Cluster → 
FASSE Cluster → 
Cannon Cluster → 
login.rc.fas.harvard.edu →
    Dec 8, 11:00:00 GMT+0 - Identified - Monthly maintenance will take place on December 8th. Our maintenance tasks should be completed between 9am-1pm. However: 

_Additionally_, MGHPCC will be performing power upgrades on the odd side of Row 8A where much of our computer resides. This is the final upgrade for this row. Current estimate for this work is a 12 hour window 6am-6pm.

A list of the affected partitions is provided at the bottom of this notice. The nodes in those partitions will be drained prior to the work and will be powered down. Once the work is completed, those nodes will be returned to service. 

**Notices:**

* New FASSE partition `fasse_gpu_h200`. This partitions has 2 H200 nodes and a 3day limit. It is available now.
* 11/26 - 11/28 are university holidays (Thanksgiving). No on-site support, FASRC staff will return on 12/1.
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**  
FASSE cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**

* Power work on Row 8A odd  
   * Audience: Users of the partitions listed below  
   * Impact: These nodes and partitions will be fully or partially down all day
* OneFS (Isilon) upgrade  
   * Audience: All Isilon (Tier 1) shares  
   * Impact: Some VMs will be impacted including Cannon OOD, CBScentral, MCZapps/MCZbase, Portal, and Rclic1 (license server)
* Slurm upgrade to 25.05.5  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* Login node reboots  
   * Audience: All login node users  
   * Impact: Login nodes will reboot during the maintenance window

**Impacted Cannon Partitions (Full or Partial Outage):**

* arguelles\_delgado\_gpu\_a100
* arguelles\_delgado\_gpu\_mixed
* bigmem\_intermediate
* blackhole\_gpu
* eddy
* gershman
* gpu\_requeue
* hejazi
* hernquist\_ice
* hoekstra
* huce\_ice
* iaifi\_gpu
* iaifi\_gpu\_priority
* iaifi\_gpu\_requeue
* itc\_gpu
* jshapiro
* kempner
* kempner\_dev
* kempner\_priority
* kempner\_h100
* kempner\_h100\_priority
* kempner\_h100\_priority2
* kempner\_h100\_priority3
* kempner\_interactive
* kempner\_requeue
* kovac
* kozinsky
* kozinsky\_gpu
* kozinsky\_priority
* kozinsky\_requeue
* murphy\_ice
* ortegahernandez\_ice
* rivas
* seas\_compute
* seas\_gpu
* serial\_requeue
* siag\_combo
* siag\_gpu
* sur
* zhuang Dec 8, 11:00:01 GMT+0 - Identified - Maintenance is now in progress Dec 8, 23:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 12 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Monthly maintenance will take place on December 8th. Our maintenance tasks should be completed between 9am-1pm. However: 

_Additionally_, MGHPCC will be performing power upgrades on the odd side of Row 8A where much of our computer resides. This is the final upgrade for this row. Current estimate for this work is a 12 hour window 6am-6pm.

A list of the affected partitions is provided at the bottom of this notice. The nodes in those partitions will be drained prior to the work and will be powered down. Once the work is completed, those nodes will be returned to service. 

**Notices:**

* New FASSE partition `fasse_gpu_h200`. This partitions has 2 H200 nodes and a 3day limit. It is available now.
* 11/26 - 11/28 are university holidays (Thanksgiving). No on-site support, FASRC staff will return on 12/1.
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* We&#039;d love to hear success stories about your or your lab&#039;s use of FASRC. Submit your story [here](https://www.rc.fas.harvard.edu/user-stories/).

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**  
FASSE cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**

* Power work on Row 8A odd  
   * Audience: Users of the partitions listed below  
   * Impact: These nodes and partitions will be fully or partially down all day
* OneFS (Isilon) upgrade  
   * Audience: All Isilon (Tier 1) shares  
   * Impact: Some VMs will be impacted including Cannon OOD, CBScentral, MCZapps/MCZbase, Portal, and Rclic1 (license server)
* Slurm upgrade to 25.05.5  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* Login node reboots  
   * Audience: All login node users  
   * Impact: Login nodes will reboot during the maintenance window

**Impacted Cannon Partitions (Full or Partial Outage):**

* arguelles\_delgado\_gpu\_a100
* arguelles\_delgado\_gpu\_mixed
* bigmem\_intermediate
* blackhole\_gpu
* eddy
* gershman
* gpu\_requeue
* hejazi
* hernquist\_ice
* hoekstra
* huce\_ice
* iaifi\_gpu
* iaifi\_gpu\_priority
* iaifi\_gpu\_requeue
* itc\_gpu
* jshapiro
* kempner
* kempner\_dev
* kempner\_priority
* kempner\_h100
* kempner\_h100\_priority
* kempner\_h100\_priority2
* kempner\_h100\_priority3
* kempner\_interactive
* kempner\_requeue
* kovac
* kozinsky
* kozinsky\_gpu
* kozinsky\_priority
* kozinsky\_requeue
* murphy\_ice
* ortegahernandez\_ice
* rivas
* seas\_compute
* seas\_gpu
* serial\_requeue
* siag\_combo
* siag\_gpu
* sur
* zhuang.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;23:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 8 Dec 2025 11:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmi4yi1m400u3o9chvqic8p37</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmi4yi1m400u3o9chvqic8p37</guid>
</item>

<item>
  <title>holylfs04 migrations</title>
  <description>
    Type: Maintenance
    Duration: 4 days, 1 hour and 10 minutes

    Affected Components: HolyLFS04 (Tier 0)
    Dec 5, 14:00:00 GMT+0 - Identified - The holylfs04 migration to holylfs06 has begun. All holylfs04 folders will be **read-only** for the duration of the migration, from **Friday, December 5th at 9AM until end of day on Monday, December 8th.** 

All labs with holylfs04 have been informed via email; please email [rdm@rc.fas.harvard.edu](mailto:rdm@rc.fas.harvard.edu) if you have any questions. Dec 9, 15:09:30 GMT+0 - Completed - Maintenance has completed successfully. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 days, 1 hour and 10 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 5&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  The holylfs04 migration to holylfs06 has begun. All holylfs04 folders will be **read-only** for the duration of the migration, from **Friday, December 5th at 9AM until end of day on Monday, December 8th.** 

All labs with holylfs04 have been informed via email; please email [rdm@rc.fas.harvard.edu](mailto:rdm@rc.fas.harvard.edu) if you have any questions..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 9&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;15:09:30&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 5 Dec 2025 14:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmisykt3v0ayak7rmnsl4btnt</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmisykt3v0ayak7rmnsl4btnt</guid>
</item>

<item>
  <title>NESE tape system maintenance 12/1/25-12/5/25</title>
  <description>
    Type: Maintenance
    Duration: 5 days

    Affected Components: NESE (NorthEast Storage Exchange), Tape - (Tier 3)
    Dec 1, 11:00:00 GMT+0 - Identified - NESE, the Northeast Storage Exchange at MGHPCC which supplies the Tier3 tape service used by FASRC, will be offline for maintenance on the system Dec 1st - 5th. There will be ongoing performance-affecting maintenance until Dec 12th. Please see below for details.

WHO: Any lab who has or is moving data to tape.

IMPACT: No access 12/1/25 - 12/5/25\. Reduced performance 12/5/25 - 12/12/25.

&gt; NESE tape system maintenance and major software upgrade is scheduled to begin on December 1, 2025\. As a result, the NESE Tape service will be offline from December 1 to December 5.
&gt; 
&gt; Starting December 8 through December 12, the service will be back online with reduced performance. All maintenance activities are planned to conclude on December 12, 2025.
&gt; 
&gt; * Monitor: &lt;https://nese.instatus.com/&gt; for real-time updates on progress
&gt; * Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements Dec 1, 11:00:01 GMT+0 - Identified - Maintenance is now in progress Dec 6, 11:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 5 days</p>
    <p><strong>Affected Components:</strong> , </p>
    &lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 1&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  NESE, the Northeast Storage Exchange at MGHPCC which supplies the Tier3 tape service used by FASRC, will be offline for maintenance on the system Dec 1st - 5th. There will be ongoing performance-affecting maintenance until Dec 12th. Please see below for details.

WHO: Any lab who has or is moving data to tape.

IMPACT: No access 12/1/25 - 12/5/25\. Reduced performance 12/5/25 - 12/12/25.

&gt; NESE tape system maintenance and major software upgrade is scheduled to begin on December 1, 2025\. As a result, the NESE Tape service will be offline from December 1 to December 5.
&gt; 
&gt; Starting December 8 through December 12, the service will be back online with reduced performance. All maintenance activities are planned to conclude on December 12, 2025.
&gt; 
&gt; * Monitor: &lt;https://nese.instatus.com/&gt; for real-time updates on progress
&gt; * Subscribe to &lt;https://nese.instatus.com/subscribe/email&gt; for updates and announcements.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 1&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Dec &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 1 Dec 2025 11:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmhjab8ow00xbfmtmh0g1qmos</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmhjab8ow00xbfmtmh0g1qmos</guid>
</item>

<item>
  <title>Starfish dashboard maintenance Nov. 14th 5-6PM</title>
  <description>
    Type: Maintenance
    Duration: 1 hour

    Affected Components: Starfish
    Nov 14, 22:00:01 GMT+0 - Identified - Maintenance is now in progress Nov 14, 23:00:00 GMT+0 - Completed - Maintenance has completed successfully Nov 14, 22:00:00 GMT+0 - Identified - There is a planned upgrade of the Starfish dashboard scheduled for Friday November 14th starting at 5PM.   
The dashboard will be down for an hour while the upgrade is performed. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 1 hour</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 14&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;22:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 14&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;23:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 14&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;22:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  There is a planned upgrade of the Starfish dashboard scheduled for Friday November 14th starting at 5PM.   
The dashboard will be down for an hour while the upgrade is performed..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Fri, 14 Nov 2025 22:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmhnyg8g701g4yv5cvuwz2hjz</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmhnyg8g701g4yv5cvuwz2hjz</guid>
</item>

<item>
  <title>Monthly Maintenance and MGHPCC Power Work - Nov. 3, 2025 6am-6pm</title>
  <description>
    Type: Maintenance
    Duration: 9 hours and 4 minutes

    Affected Components: , , SLURM Scheduler - FASSE, , Kempner Cluster CPU, Cannon Compute Cluster (Holyoke), , Cannon Open OnDemand/VDI, SLURM Scheduler - Cannon, FASSE Open OnDemand/VDI, seas_compute, Kempner Cluster GPU, Login Nodes - Holyoke, , GPU nodes (Holyoke), FASSE Compute Cluster (Holyoke), FASSE login nodes, Login Nodes - Boston, Netscratch (Global Scratch), Boston Compute Nodes, 
Login Nodes → 
VDI/OpenOnDemand → 
Kempner Cluster → 
FASSE Cluster → 
Cannon Cluster →
    Nov 3, 11:00:00 GMT+0 - Identified - Monthly maintenance will take place on November 3rd. Additionally, MGHPCC will be performing power upgrades on the even side of Row 8A where much of our computer resides. A further upgrade will take place Dec. 8th on the odd side.

A list of the affected partitions is provided at the bottom of this notice. The nodes in those partitions will be drained prior to the work and will be powered down. Once the work is completed, those nodes will be returned to service. Current estimate is a 12 hour window. We will adjust as we know more.

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**  
FASSE cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**

* Power work on Row 8A Even  
   * Audience: Users of the partitions listed below  
   * Impact: These nodes and partitions will be fully or partially down all day
* Slurm upgrade to 25.05.4  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* Block repo.anaconda.com cluster wide  
   * Audience: Anyone attempting to use repo.anaconda.com  
   * Impact: This change should not impact your Python workflow on the cluster. But if it does, consider using the open-source channel, `conda-forge`, through Miniforge distribution to install Python packages. This can be done by following our instructions on &lt;https://docs.rc.fas.harvard.edu/kb/python-package-installation/&gt;
* Change Slurm User to Local User  
   * Audience: All cluster users  
   * Impact: Behind the scenes. No impact to users
* Login node reboots (morning)  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

**AFFECTED PARTITIONS** 
Nov. 3, 2025 - All Day Power Work  
Partial or Full Outage Apples to:

arguelles\_delgado\_h100

bigmem

dvorkin

eddy

enos

gpu

gpu\_h200

gpu\_requeue

hsph

hsph\_gpu

intermediate

itc\_cluster

joonholee

jshapiro

kempner\_dev

kemkpner\_eng

kempner\_requeue

mweber\_compute

mweber\_gpu

olveczky\_sapphire

sapphire

seas\_compute

seas\_gpu

serial\_requeue

yao

yao\_gpu

yao\_priority

test Nov 3, 11:00:01 GMT+0 - Identified - Maintenance is now in progress Nov 3, 20:04:13 GMT+0 - Completed - Maintenance has completed successfully including power work at MGHPCC.

A reminder that additional all-day power work will take place on Dec 8th, along with our maintenance from 9am-1pm 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 9 hours and 4 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 3&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Monthly maintenance will take place on November 3rd. Additionally, MGHPCC will be performing power upgrades on the even side of Row 8A where much of our computer resides. A further upgrade will take place Dec. 8th on the odd side.

A list of the affected partitions is provided at the bottom of this notice. The nodes in those partitions will be drained prior to the work and will be powered down. Once the work is completed, those nodes will be returned to service. Current estimate is a 12 hour window. We will adjust as we know more.

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**  
FASSE cluster will be paused during this maintenance?: **PARTIAL OUTAGE/YES**

* Power work on Row 8A Even  
   * Audience: Users of the partitions listed below  
   * Impact: These nodes and partitions will be fully or partially down all day
* Slurm upgrade to 25.05.4  
   * Audience: All cluster users  
   * Impact: Jobs will be paused during maintenance
* Block repo.anaconda.com cluster wide  
   * Audience: Anyone attempting to use repo.anaconda.com  
   * Impact: This change should not impact your Python workflow on the cluster. But if it does, consider using the open-source channel, `conda-forge`, through Miniforge distribution to install Python packages. This can be done by following our instructions on &lt;https://docs.rc.fas.harvard.edu/kb/python-package-installation/&gt;
* Change Slurm User to Local User  
   * Audience: All cluster users  
   * Impact: Behind the scenes. No impact to users
* Login node reboots (morning)  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

**AFFECTED PARTITIONS** 
Nov. 3, 2025 - All Day Power Work  
Partial or Full Outage Apples to:

arguelles\_delgado\_h100

bigmem

dvorkin

eddy

enos

gpu

gpu\_h200

gpu\_requeue

hsph

hsph\_gpu

intermediate

itc\_cluster

joonholee

jshapiro

kempner\_dev

kemkpner\_eng

kempner\_requeue

mweber\_compute

mweber\_gpu

olveczky\_sapphire

sapphire

seas\_compute

seas\_gpu

serial\_requeue

yao

yao\_gpu

yao\_priority

test.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 3&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Nov &lt;var data-var=&#039;date&#039;&gt; 3&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;20:04:13&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully including power work at MGHPCC.

A reminder that additional all-day power work will take place on Dec 8th, along with our maintenance from 9am-1pm.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 3 Nov 2025 11:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmggy7q9801vdtke67ycm4dxq</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmggy7q9801vdtke67ycm4dxq</guid>
</item>

<item>
  <title>FASRC monthly maintenance Monday October 6th, 2025 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: , Network - Cambridge, Network - Boston, Network - Holyoke/MGHPCC, Login Nodes - Boston, Netscratch (Global Scratch), Login Nodes - Holyoke, FASSE login nodes, 
Login Nodes →
    Oct 6, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Oct 6, 13:00:00 GMT+0 - Identified - FASRC monthly maintenance will take place Monday October 6th, 2025 from 9am-1pm

**NOTICES**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays: Columbus / Indigenous Peoples’ Day - October 13

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* DNS server reboots  
   * Audience: All FASRC services  
   * Impact: Rolling reboot should have no impact
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt; Oct 6, 17:00:00 GMT+0 - Completed - Maintenance is now in progress 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Oct &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Oct &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  FASRC monthly maintenance will take place Monday October 6th, 2025 from 9am-1pm

**NOTICES**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays: Columbus / Indigenous Peoples’ Day - October 13

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* DNS server reboots  
   * Audience: All FASRC services  
   * Impact: Rolling reboot should have no impact
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt;.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Oct &lt;var data-var=&#039;date&#039;&gt; 6&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 6 Oct 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmg1cda8s098z4gs4ylropic7</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmg1cda8s098z4gs4ylropic7</guid>
</item>

<item>
  <title>VPN concentrator rolling updates overnight</title>
  <description>
    Type: Maintenance
    

    
    Sep 30, 07:00:00 GMT+0 - Completed - Maintenance has completed successfully Sep 30, 06:00:00 GMT+0 - Completed - Networking will be patching the VPN concentrators overnight. This will be done in a rolling order so that one ore more are always online.   
  
This may cause active VPN connections to drop, but they can be re-connected shortly after. ETA is one hour total. Sep 30, 06:00:00 GMT+0 - Identified - Networking will be patching the VPN concentrators overnight. This will be done in a rolling order so that one ore more are always online.   
  
This may cause active VPN connections to drop, but they can be re-connected shortly after. ETA is one hour total. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    
    
    &lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 30&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;07:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 30&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;06:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Networking will be patching the VPN concentrators overnight. This will be done in a rolling order so that one ore more are always online.   
  
This may cause active VPN connections to drop, but they can be re-connected shortly after. ETA is one hour total..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 30&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;06:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Networking will be patching the VPN concentrators overnight. This will be done in a rolling order so that one ore more are always online.   
  
This may cause active VPN connections to drop, but they can be re-connected shortly after. ETA is one hour total..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Tue, 30 Sep 2025 06:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmg5awjiw04n710wcsl3zi0gv</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmg5awjiw04n710wcsl3zi0gv</guid>
</item>

<item>
  <title>FASRC monthly maintenance Monday September 8th, 2025 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: Kempner Cluster CPU, Kempner Cluster GPU, , Cannon Compute Cluster (Holyoke), Boston Compute Nodes, Netscratch (Global Scratch), , , SLURM Scheduler - FASSE, , GPU nodes (Holyoke), Login Nodes - Boston, FASSE login nodes, Login Nodes - Holyoke, , FASSE Open OnDemand/VDI, Cannon Open OnDemand/VDI, SLURM Scheduler - Cannon, FASSE Compute Cluster (Holyoke), seas_compute, 
Login Nodes → 
Cannon Cluster → 
FASSE Cluster → 
VDI/OpenOnDemand → 
Kempner Cluster →
    Sep 8, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Sep 8, 17:00:00 GMT+0 - Completed - Maintenance has completed successfully Sep 8, 13:00:00 GMT+0 - Identified - FASRC monthly maintenance will take place Monday September 8th, 2025 from 9am-1pm

**NOTICES**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays: Labor Day, Monday September 1st

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm Upgrade to 25.05.2  
   * Audience: All cluster users  
   * Impact: Jobs and the scheduler will be paused during this upgrade
* Domain controller work  
   * Audience: Internal network  
   * Impact: No impact expected
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt; 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Sep &lt;var data-var=&#039;date&#039;&gt; 8&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  FASRC monthly maintenance will take place Monday September 8th, 2025 from 9am-1pm

**NOTICES**

* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays: Labor Day, Monday September 1st

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm Upgrade to 25.05.2  
   * Audience: All cluster users  
   * Impact: Jobs and the scheduler will be paused during this upgrade
* Domain controller work  
   * Audience: Internal network  
   * Impact: No impact expected
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will rebooted during this maintenance window
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt;.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 8 Sep 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmevo0dad0012rwy6qv5mv34j</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmevo0dad0012rwy6qv5mv34j</guid>
</item>

<item>
  <title>Starfish upgrade Wednesday, August 20th 5PM-7PM</title>
  <description>
    Type: Maintenance
    Duration: 2 hours

    Affected Components: Starfish
    Aug 20, 21:00:01 GMT+0 - Identified - Maintenance is now in progress Aug 20, 21:00:00 GMT+0 - Identified - Starfish will be performing an upgrade on Wednesday, August 20th from 5PM-7PM. The web interface will be unavailable during that timeframe. Aug 20, 23:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 2 hours</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 20&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 20&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Starfish will be performing an upgrade on Wednesday, August 20th from 5PM-7PM. The web interface will be unavailable during that timeframe..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 20&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;23:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Wed, 20 Aug 2025 21:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmehjmf9q000dam879fafq3vq</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmehjmf9q000dam879fafq3vq</guid>
</item>

<item>
  <title>SEAS: seas_gpu partition GPU upgrades 8/11 - 8/14</title>
  <description>
    Type: Maintenance
    Duration: 4 days, 9 hours and 39 minutes

    Affected Components: seas_compute
    Aug 11, 10:00:00 GMT+0 - Identified - Between 8/11/25 6AM - 8/14/25 5PM FASRC will be upgrading 14 of the H100 GPU nodes in the `seas_gpu `partition to H200 GPUs. This sill also affect `mweber_gpu`

A reservation has been set which will drain the nodes of jobs prior to the maintenance. The SEAS GPU partition will be running at 75% capacity during these updates. FASRC has hundreds of GPU, so users should feel free to utilize `gpu_requeue` if needed for their jobs.

Affected nodes: 

`mweber_gpu` nodes (13):

```
holygpu8a[18204,18301-18304,18401-18404,18501-18502,18601-18602]
```

seas\_gpu nodes (14):

```
holygpu8a[16101-16104,16201-16204,16301-16304,16401-16402]
```

Please reach out to [rchelp@rc.fas.harvard.edu](mailto:rchelp@rc.fas.harvard.edu) if you have any questions or concerns. Aug 15, 14:13:24 GMT+0 - Identified - Maintenance is still in progress - imaging of the new H200 nodes is ongoing. Current ETA is end of day Friday. For further questions, please contact rchelp@rc.fas.harvard.edu Aug 15, 19:39:27 GMT+0 - Completed - Maintenance has completed successfully. SEAS H200 nodes have been imaged and are back in service.  Aug 11, 10:00:01 GMT+0 - Identified - Maintenance is now in progress 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 days, 9 hours and 39 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 11&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;10:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Between 8/11/25 6AM - 8/14/25 5PM FASRC will be upgrading 14 of the H100 GPU nodes in the `seas_gpu `partition to H200 GPUs. This sill also affect `mweber_gpu`

A reservation has been set which will drain the nodes of jobs prior to the maintenance. The SEAS GPU partition will be running at 75% capacity during these updates. FASRC has hundreds of GPU, so users should feel free to utilize `gpu_requeue` if needed for their jobs.

Affected nodes: 

`mweber_gpu` nodes (13):

```
holygpu8a[18204,18301-18304,18401-18404,18501-18502,18601-18602]
```

seas\_gpu nodes (14):

```
holygpu8a[16101-16104,16201-16204,16301-16304,16401-16402]
```

Please reach out to [rchelp@rc.fas.harvard.edu](mailto:rchelp@rc.fas.harvard.edu) if you have any questions or concerns..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 15&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:13:24&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is still in progress - imaging of the new H200 nodes is ongoing. Current ETA is end of day Friday. For further questions, please contact rchelp@rc.fas.harvard.edu.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 15&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;19:39:27&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully. SEAS H200 nodes have been imaged and are back in service. .&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 11&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;10:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 11 Aug 2025 10:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cme1un2le0626okcc24iokqze</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cme1un2le0626okcc24iokqze</guid>
</item>

<item>
  <title>Monthly maintenance August 4, 2025 9am-1pm</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: , Login Nodes - Boston, Netscratch (Global Scratch), FASSE login nodes, Login Nodes - Holyoke, 
Login Nodes →
    Aug 4, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Aug 4, 17:00:00 GMT+0 - Completed - Maintenance has completed successfully Aug 4, 13:00:00 GMT+0 - Identified - &gt; _Important note about changes to the gpu\_test, test, and remoteviz partitions_
&gt; 
&gt; On Monday July 28th we will make the following changes. These changes are necessary in order to reduce congestion on these partitions.
&gt; 
&gt; gpu\_test will have a 2 job limit per user - Reminder, these partitions are for testing and debugging, not for production work.
&gt; 
&gt; gpu\_test, test, and remoteviz partitions will no longer be available for multi-partition submission

FASRC monthly maintenance will take place Monday August 4th, 2025 from 9am-1pm

**MONTHLY NOTICES**

* Do you have a success story about your use of the FASRC clusters or services? We&#039;d love to hear it and post it on our [new User Stories page.](https://www.rc.fas.harvard.edu/user-stories/)
* New Quota tool available. Type quota -h to see the full instructions for usage or visit the usage doc.
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* Please see note above about partition changes  
   * Those changes will happen on July 28th
* FASRC websites (www.rc and docs.rc) will be updated  
   * Audience: Anyone browsing our websites  
   * Impact: The sites will be periodically unavailable during reboots
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will reboot during this maintenance window
* Netscratch cleanup (&lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt;)  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed.

Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,

FAS Research Computing

&lt;https://docs.rc.fas.harvard.edu/&gt;

&lt;https://www.rc.fas.harvard.edu/&gt; 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , </p>
    &lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 4&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 4&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Aug &lt;var data-var=&#039;date&#039;&gt; 4&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  &gt; _Important note about changes to the gpu\_test, test, and remoteviz partitions_
&gt; 
&gt; On Monday July 28th we will make the following changes. These changes are necessary in order to reduce congestion on these partitions.
&gt; 
&gt; gpu\_test will have a 2 job limit per user - Reminder, these partitions are for testing and debugging, not for production work.
&gt; 
&gt; gpu\_test, test, and remoteviz partitions will no longer be available for multi-partition submission

FASRC monthly maintenance will take place Monday August 4th, 2025 from 9am-1pm

**MONTHLY NOTICES**

* Do you have a success story about your use of the FASRC clusters or services? We&#039;d love to hear it and post it on our [new User Stories page.](https://www.rc.fas.harvard.edu/user-stories/)
* New Quota tool available. Type quota -h to see the full instructions for usage or visit the usage doc.
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;

**MAINTENANCE TASKS**

Cannon cluster will be paused during this maintenance?: **NO**  
FASSE cluster will be paused during this maintenance?: **NO**

* Please see note above about partition changes  
   * Those changes will happen on July 28th
* FASRC websites (www.rc and docs.rc) will be updated  
   * Audience: Anyone browsing our websites  
   * Impact: The sites will be periodically unavailable during reboots
* Login node reboots  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will reboot during this maintenance window
* Netscratch cleanup (&lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt;)  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed.

Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,

FAS Research Computing

&lt;https://docs.rc.fas.harvard.edu/&gt;

&lt;https://www.rc.fas.harvard.edu/&gt;.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 4 Aug 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmdrj0l39001qxxr70vtdmbos</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmdrj0l39001qxxr70vtdmbos</guid>
</item>

<item>
  <title>FASRC away for all-hands meeting 7/17</title>
  <description>
    Type: Maintenance
    Duration: 8 hours

    Affected Components: FASRC Ticket System (ServiceNow)
    Jul 17, 13:00:01 GMT+0 - Identified - Maintenance is now in progress Jul 17, 13:00:00 GMT+0 - Identified - FASRC staff will be attending an all-hands meeting all day Thursday 7/17/25.

Ticket response will be delayed. Jul 17, 21:00:00 GMT+0 - Completed - Maintenance has completed successfully 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 8 hours</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 17&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 17&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  FASRC staff will be attending an all-hands meeting all day Thursday 7/17/25.

Ticket response will be delayed..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 17&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 17 Jul 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmd530cd302i63cj7eggitvvl</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmd530cd302i63cj7eggitvvl</guid>
</item>

<item>
  <title>FASRC Monthly maintenance July 7, 2025 9AM-1PM</title>
  <description>
    Type: Maintenance
    Duration: 4 hours

    Affected Components: , , , GPU nodes (Holyoke), Cannon Open OnDemand/VDI, SLURM Scheduler - Cannon, FASSE login nodes, seas_compute, FASSE Open OnDemand/VDI, Kempner Cluster CPU, , Cannon Compute Cluster (Holyoke), FASSE Compute Cluster (Holyoke), Login Nodes - Holyoke, Kempner Cluster GPU, Netscratch (Global Scratch), , SLURM Scheduler - FASSE, Boston Compute Nodes, Login Nodes - Boston, 
Kempner Cluster → 
Cannon Cluster → 
VDI/OpenOnDemand → 
FASSE Cluster → 
Login Nodes →
    Jul 7, 17:00:00 GMT+0 - Completed - Maintenance has completed successfully Jul 7, 13:00:00 GMT+0 - Identified - FASRC monthly maintenance will take place Monday July 7th, 2025 from 9am-1pm

**NOTICES**

* ​New Quota tool available (/usr/local/sbin/quota) - Works on _all_ filesystem types (home directory, lustre, isilon, netscratch, etc.)  
Type `quota -h` to see the full instructions for usage o visit [the usage doc](https://docs.rc.fas.harvard.edu/kb/checking-quota-and-usage/).
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays:​ Juneteenth - ​T​hur. June 19​ / Independence Day - Fri​. July 4

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm Upgrade to 24.11.5  
   * Audience: All cluster users  
   * Impact: Jobs and the scheduler will be paused during this upgrade
* Login node ​OS ​upgrades  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will ​upgraded ​and unavailable during this maintenance window
* ​Start of cluster OS upgrades - July 7 -10  
   * Audience: All cluster users  
   * Impact: **Over 4 days,** July 7 through 10, we will upgrade the OS on 25% of the cluster each day. During that time, total capacity will be reduced across the cluster by 1/4 each day. This will require draining each sub-set of nodes ahead of time.
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt; Jul 7, 13:00:01 GMT+0 - Identified - Maintenance is now in progress 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 hours</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;17:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  FASRC monthly maintenance will take place Monday July 7th, 2025 from 9am-1pm

**NOTICES**

* ​New Quota tool available (/usr/local/sbin/quota) - Works on _all_ filesystem types (home directory, lustre, isilon, netscratch, etc.)  
Type `quota -h` to see the full instructions for usage o visit [the usage doc](https://docs.rc.fas.harvard.edu/kb/checking-quota-and-usage/).
* Training: Upcoming training from FASRC and other sources can be found on our Training Calendar. at &lt;https://www.rc.fas.harvard.edu/upcoming-training/&gt;
* Status Page: You can subscribe to our status to receive notifications of maintenance, incidents, and their resolution at &lt;https://status.rc.fas.harvard.edu/&gt; (click Get Updates for options).
* Upcoming holidays:​ Juneteenth - ​T​hur. June 19​ / Independence Day - Fri​. July 4

**MAINTENANCE TASKS**  
Cannon cluster will be paused during this maintenance?: **YES**  
FASSE cluster will be paused during this maintenance?: **YES**

* Slurm Upgrade to 24.11.5  
   * Audience: All cluster users  
   * Impact: Jobs and the scheduler will be paused during this upgrade
* Login node ​OS ​upgrades  
   * Audience: Anyone logged into a FASRC Cannon or FASSE login node  
   * Impact: All login nodes will ​upgraded ​and unavailable during this maintenance window
* ​Start of cluster OS upgrades - July 7 -10  
   * Audience: All cluster users  
   * Impact: **Over 4 days,** July 7 through 10, we will upgrade the OS on 25% of the cluster each day. During that time, total capacity will be reduced across the cluster by 1/4 each day. This will require draining each sub-set of nodes ahead of time.
* Netscratch cleanup ( &lt;https://docs.rc.fas.harvard.edu/kb/policy-scratch/&gt; )  
   * Audience: Cluster users  
   * Impact: Files older than 90 days will be removed. Please note that retention cleanup can and does run at any time, not just during the maintenance window.

Thank you,  
FAS Research Computing  
&lt;https://docs.rc.fas.harvard.edu/&gt;  
&lt;https://www.rc.fas.harvard.edu/&gt;.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 7 Jul 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmc0yfhrq0012c3y9c28s35se</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmc0yfhrq0012c3y9c28s35se</guid>
</item>

<item>
  <title>Rolling cluster OS upgrades July 7 - 10</title>
  <description>
    Type: Maintenance
    Duration: 4 days, 3 hours and 3 minutes

    Affected Components: , FASSE Open OnDemand/VDI, , GPU nodes (Holyoke), Cannon Open OnDemand/VDI, Kempner Cluster CPU, Kempner Cluster GPU, seas_compute, FASSE Compute Cluster (Holyoke), Cannon Compute Cluster (Holyoke), Boston Compute Nodes, , 
FASSE Cluster → 
VDI/OpenOnDemand → 
Cannon Cluster →
    Jul 7, 13:00:01 GMT+0 - Identified - Cannon rolling upgrades are in progress. Not all nodes are available.

&lt;https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/&gt; Jul 7, 13:00:01 GMT+0 - Identified - **UPDATE:** 7/7/25 6M FASSE is operational.

~~Please be aware that FASSE jobs cannot be launched at this time due to the upgrades.~~ 
~~We will return all FASSE nodes to normal services as soon as possible.~~

&lt;https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/&gt; Jul 7, 13:00:00 GMT+0 - Identified - Cluster OS upgrades - July 7 -10

* Audience: All cluster users
* Impact: **Over 4 days**, July 7 through 10, we will upgrade the OS on **25%** of the cluster each day.  
During that time, total capacity will be reduced across the cluster by 1/4 each day.  
This will require draining each sub-set of nodes ahead of time.

Work begins during the July 7th maintenance (login nogdes will be upgraded during the 7/7 maintenance window) and will continue through July 10th.

Additional details and a breakdown of each phase: [2025 Compute OS Upgrade](https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/) Jul 11, 16:02:45 GMT+0 - Completed - All upgrades are complete. A small number of nodes need clean-up, but the cluster is back to normal operation with all nodes running Rocky 8.10\. Thanks for your patience. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 4 days, 3 hours and 3 minutes</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Cannon rolling upgrades are in progress. Not all nodes are available.

&lt;https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/&gt;.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  **UPDATE:** 7/7/25 6M FASSE is operational.

~~Please be aware that FASSE jobs cannot be launched at this time due to the upgrades.~~ 
~~We will return all FASSE nodes to normal services as soon as possible.~~

&lt;https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/&gt;.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 7&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Cluster OS upgrades - July 7 -10

* Audience: All cluster users
* Impact: **Over 4 days**, July 7 through 10, we will upgrade the OS on **25%** of the cluster each day.  
During that time, total capacity will be reduced across the cluster by 1/4 each day.  
This will require draining each sub-set of nodes ahead of time.

Work begins during the July 7th maintenance (login nogdes will be upgraded during the 7/7 maintenance window) and will continue through July 10th.

Additional details and a breakdown of each phase: [2025 Compute OS Upgrade](https://www.rc.fas.harvard.edu/blog/2025-compute-os-upgrade/).&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jul &lt;var data-var=&#039;date&#039;&gt; 11&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;16:02:45&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  All upgrades are complete. A small number of nodes need clean-up, but the cluster is back to normal operation with all nodes running Rocky 8.10\. Thanks for your patience..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 7 Jul 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmc0yoo9m001qc3y986idbaqt</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmc0yoo9m001qc3y986idbaqt</guid>
</item>

<item>
  <title>June 5-6 MGHPCC pod 7c cooling updates - See partition list below</title>
  <description>
    Type: Maintenance
    Duration: 8 hours and 3 minutes

    Affected Components: seas_compute, Cannon Compute Cluster (Holyoke)
    Jun 5, 11:00:01 GMT+0 - Identified - Maintenance is now in progress Jun 5, 19:03:03 GMT+0 - Completed - The work on row 7c is complete. Returning idled nodes to normal service. Jun 5, 11:00:00 GMT+0 - Identified - There will be additional scheduled maintenance at MGHPCC between June 5th and 6th. 

As part of the work during the MGHPCC Outage, one of the Cooling Distribution Unit (CDU) in Pod 7c will be replaced. This will allow for future expansion into this space. 

This work will run from Thursday Jun 5th until the evening of Friday June 6th. This means nodes whose names begin with holy7c02, 04, 06, 08, 10, 12 _will not_ come back online after the outage and will remain down until this CDU update is complete.

This impacts the following partitions. If you are using one of those partitions please use the public sapphire partition while your equipment is being serviced. These nodes will be returned to service once the CDU work is complete: 

* blackhole
* blackhole\_priority
* davies
* desai
* eddy
* huce\_cascade
* huce\_cascade\_priority
* huttenhower
* jacobsen2
* janson
* janson\_cascade
* ke
* lukin
* nguyen
* seas\_compute
* shared
* tambe
* vishwanath
* whipple
* xlin 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 8 hours and 3 minutes</p>
    <p><strong>Affected Components:</strong> , </p>
    &lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 5&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 5&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;19:03:03&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  The work on row 7c is complete. Returning idled nodes to normal service..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 5&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;11:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  There will be additional scheduled maintenance at MGHPCC between June 5th and 6th. 

As part of the work during the MGHPCC Outage, one of the Cooling Distribution Unit (CDU) in Pod 7c will be replaced. This will allow for future expansion into this space. 

This work will run from Thursday Jun 5th until the evening of Friday June 6th. This means nodes whose names begin with holy7c02, 04, 06, 08, 10, 12 _will not_ come back online after the outage and will remain down until this CDU update is complete.

This impacts the following partitions. If you are using one of those partitions please use the public sapphire partition while your equipment is being serviced. These nodes will be returned to service once the CDU work is complete: 

* blackhole
* blackhole\_priority
* davies
* desai
* eddy
* huce\_cascade
* huce\_cascade\_priority
* huttenhower
* jacobsen2
* janson
* janson\_cascade
* ke
* lukin
* nguyen
* seas\_compute
* shared
* tambe
* vishwanath
* whipple
* xlin.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 5 Jun 2025 11:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmae54uxs002g2c3ys3k5b51r</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmae54uxs002g2c3ys3k5b51r</guid>
</item>

<item>
  <title>2025 MGHPCC power downtime June 2-4, 2025 </title>
  <description>
    Type: Maintenance
    Duration: 3 days

    Affected Components: SLURM Scheduler - Cannon, Holyoke/MGHPCC Data Center, Software &amp; Modules, NESE (NorthEast Storage Exchange), Holystore01 (Tier 0), FASSE login nodes, FASSE Open OnDemand/VDI, seas_compute, Kempner Cluster CPU, Isilon Storage Holyoke (Tier 1), Holyoke Firewall, Holyoke Tier 2 NFS (new), HolyLFS05 (Tier 0), Globus Data Transfer, Netscratch (Global Scratch), GPU nodes (Holyoke), Infiniband - Holyoke/MGHPCC, HolyLFS04 (Tier 0), SLURM Scheduler - FASSE, Boston Compute Nodes, Samba Cluster, Virtual Infrastructure - Holyoke, Network - Holyoke/MGHPCC, License Servers, Cannon Compute Cluster (Holyoke), FASSE Compute Cluster (Holyoke), Holylabs, Cannon Open OnDemand/VDI, Login Nodes - Holyoke, Kempner Cluster GPU, holECS, Holyoke Specialty Storage, Login Nodes - Boston, HolyLFS06 (Tier 0), Tape - (Tier 3)
    Jun 2, 13:00:00 GMT+0 - Identified - The yearly power downtime at our Holyoke data center, MGHPCC, has been scheduled.   
This year&#039;s power downtime will take place on Tuesday June 3, 2025\. 

This will require FASRC to begin shutdown of our systems beginning at _9AM on Monday, June 2nd_.   
We have worked to reduce the total outage time this year.  
We will begin power-up on Wednesday June 4th with an expected return to full service by _9AM Thursday June 5th_.

* **Monday June 2nd** \- Power-down begins at 9AM
* **Tuesday June 3rd** \- Power out at MGHPCC
* **Wednesday June 4th** \- Maintenance tasks and then power-up begins
* **Thursday June 5th** \- Expected return to full service by 9AM

**Maintenance:**  
During this downtime, Holylabs (/n/holylabs) will move to new hardware.  
Starfish, Coldfront, and the Portal will be unavailable during the downtime.

For more details including a graphical timeline, please see: &lt;https://www.rc.fas.harvard.edu/events/2025-mghpcc-power-downtime/&gt;

**Updates will be posted here on our status page:** &lt;https://status.rc.fas.harvard.edu/&gt;   
Note that you can subscribe to receive updates as they happen. On the status page, click Get Updates.

Notices and reminders will also be sent to all users via our mailing lists. Jun 5, 13:00:00 GMT+0 - Completed - Maintenance has completed successfully Jun 2, 13:00:01 GMT+0 - Identified - Maintenance is now in progress 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 3 days</p>
    <p><strong>Affected Components:</strong> , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , </p>
    &lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  The yearly power downtime at our Holyoke data center, MGHPCC, has been scheduled.   
This year&#039;s power downtime will take place on Tuesday June 3, 2025\. 

This will require FASRC to begin shutdown of our systems beginning at _9AM on Monday, June 2nd_.   
We have worked to reduce the total outage time this year.  
We will begin power-up on Wednesday June 4th with an expected return to full service by _9AM Thursday June 5th_.

* **Monday June 2nd** \- Power-down begins at 9AM
* **Tuesday June 3rd** \- Power out at MGHPCC
* **Wednesday June 4th** \- Maintenance tasks and then power-up begins
* **Thursday June 5th** \- Expected return to full service by 9AM

**Maintenance:**  
During this downtime, Holylabs (/n/holylabs) will move to new hardware.  
Starfish, Coldfront, and the Portal will be unavailable during the downtime.

For more details including a graphical timeline, please see: &lt;https://www.rc.fas.harvard.edu/events/2025-mghpcc-power-downtime/&gt;

**Updates will be posted here on our status page:** &lt;https://status.rc.fas.harvard.edu/&gt;   
Note that you can subscribe to receive updates as they happen. On the status page, click Get Updates.

Notices and reminders will also be sent to all users via our mailing lists..&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 5&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;Jun &lt;var data-var=&#039;date&#039;&gt; 2&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;13:00:01&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Maintenance is now in progress.&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Mon, 2 Jun 2025 13:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cm8rjnw7s0004awhssij2h7j6</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cm8rjnw7s0004awhssij2h7j6</guid>
</item>

<item>
  <title>Starfish upgrade Thursday, May 29th from 5PM-6PM</title>
  <description>
    Type: Maintenance
    Duration: 17 hours and 27 minutes

    Affected Components: Starfish
    May 29, 21:00:00 GMT+0 - Identified - Starfish upgrade Thursday, May 29th from 5PM-6PM. Starfish will be unavailable during that time May 30, 14:27:29 GMT+0 - Completed - Maintenance has completed successfully. 
  </description>
  <content:encoded>
    <![CDATA[<p><strong>Type:</strong> Maintenance</p>
    <p><strong>Duration:</strong> 17 hours and 27 minutes</p>
    <p><strong>Affected Components:</strong> </p>
    &lt;p&gt;&lt;small&gt;May &lt;var data-var=&#039;date&#039;&gt; 29&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;21:00:00&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Identified&lt;/strong&gt; -
  Starfish upgrade Thursday, May 29th from 5PM-6PM. Starfish will be unavailable during that time.&lt;/p&gt;
&lt;p&gt;&lt;small&gt;May &lt;var data-var=&#039;date&#039;&gt; 30&lt;/var&gt;, &lt;var data-var=&#039;time&#039;&gt;14:27:29&lt;/var&gt; GMT+0&lt;/small&gt;&lt;br&gt;&lt;strong&gt;Completed&lt;/strong&gt; -
  Maintenance has completed successfully..&lt;/p&gt;
]]>
  </content:encoded>
  <pubDate>Thu, 29 May 2025 21:00:00 +0000</pubDate>
  <link>https://status.rc.fas.harvard.edu/maintenance/cmay295pr0023t63j9cts25br</link>
  <guid>https://status.rc.fas.harvard.edu/maintenance/cmay295pr0023t63j9cts25br</guid>
</item>

  </channel>
  </rss>