Difference between revisions of "Tier1 Operations Report 2016-12-07"
From GridPP Wiki
(Created page with "==RAL Tier1 Operations Report for 7th December 2016== __NOTOC__ ====== ====== <!-- ************************************************************* -----> <!-- ***********Start...") |
(→) |
||
Line 21: | Line 21: | ||
| style="background-color: #f8d6a9; border-bottom: 1px solid silver; text-align: center; font-size: 1em; font-weight: bold; margin-top: 0; margin-bottom: 0; padding-top: 0.1em; padding-bottom: 0.1em;" | Resolved Disk Server Issues | | style="background-color: #f8d6a9; border-bottom: 1px solid silver; text-align: center; font-size: 1em; font-weight: bold; margin-top: 0; margin-bottom: 0; padding-top: 0.1em; padding-bottom: 0.1em;" | Resolved Disk Server Issues | ||
|} | |} | ||
− | * | + | * GDSS726 (CMSDisk - D1T0) reported FSProbe errors on Thursday 1st Dec. and was taken out of service. It was returned to service the following day although the tests did not find any problems. |
+ | * GDSS747 (AtlasDataDisk - D1T0) also failed on Thursday 1st Dec. Two failed disks were found. The server was returned to service on Monday (5th Dec). | ||
+ | * GDSS650 (LHCbUser - D1T0) failed on Saturday morning, 3rd Dec. It was returned to service yesterday (6th Dec). A disk had failed - and the replacement to that disk also failed. During teh RAID rebuild a further disk drive started reporting problems and was also swapped. | ||
+ | * GDSS701 (LHCbDst - D1T0) was taken out of service on Saturday (3rd Dec) when it reported FSProbe errors when a disk was replaced. It was returned to service on the 5th Dec. | ||
<!-- ***************************************************** -----> | <!-- ***************************************************** -----> | ||
Revision as of 10:15, 7 December 2016
RAL Tier1 Operations Report for 7th December 2016
Review of Issues during the week 30th November and 7th December 2016. |
- It was found that some worker nodes were being put offline owing to clock drift. (This was around 10% of them on Monday). This was traced to a problem within the NTP daemon and was fixed by a configuration change.
Resolved Disk Server Issues |
- GDSS726 (CMSDisk - D1T0) reported FSProbe errors on Thursday 1st Dec. and was taken out of service. It was returned to service the following day although the tests did not find any problems.
- GDSS747 (AtlasDataDisk - D1T0) also failed on Thursday 1st Dec. Two failed disks were found. The server was returned to service on Monday (5th Dec).
- GDSS650 (LHCbUser - D1T0) failed on Saturday morning, 3rd Dec. It was returned to service yesterday (6th Dec). A disk had failed - and the replacement to that disk also failed. During teh RAID rebuild a further disk drive started reporting problems and was also swapped.
- GDSS701 (LHCbDst - D1T0) was taken out of service on Saturday (3rd Dec) when it reported FSProbe errors when a disk was replaced. It was returned to service on the 5th Dec.
Current operational status and issues |
- There is a problem seen by LHCb of a low but persistent rate of failure when copying the results of batch jobs to Castor. There is also a further problem that sometimes occurs when these (failed) writes are attempted to storage at other sites.
- The intermittent, low-level, load-related packet loss that has been seen over external connections is still being tracked. The replacement of the UKLight router appears to have reduced this - but we are allowing more time to pass before drawing any conclusions.
Ongoing Disk Server Issues |
- None
Notable Changes made since the last meeting. |
- Maintenance was carried out on the UPS and generator in R89 yesterday.
- There was restart test of the ECHO Ceph system yesterday> this was to understand how best to do this and set-up appropriate operating procedures.
- LHCb writing to the 'D' tapes. The migration of their data from 'C' to 'D' tapes is underway - with around 300 (some 30%) of the tapes done.
- An update to the FTS3 service (to version 3.5.7) has taken place this morning.
- Increased number of CMS multicore jobs allowed to run as the previous limit was a bit too low. (This is a further increase as compared to that of around a month ago).
Declared in the GOC DB |
None
Advanced warning for other interventions |
The following items are being discussed and are still to be formally scheduled and announced. |
Pending - but not yet formally announced:
- Firmware update on Clustervision '13 disk servers. These are distributed as follows: AtlasDataDisk: 12; CMSDisk: 5; LHCbDst: 12.
- Merge AtlasScratchDisk and LhcbUser into larger disk pools - Possible date Thursday 8th Dec.
Listing by category:
- Castor:
- Merge AtlasScratchDisk and LhcbUser into larger disk pools
- Update to Castor version 2.1.15. Planning to roll out January 2017. (Proposed dates: 10th Jan: Nameserver; 17th Jan: First stager (LHCb); 24th Jan: Stager (Atlas); 26th Jan: Stager (GEN); 31st Jan: Final stager (CMS)).
- Update SRMs to new version, including updating to SL6. This will be done after the Castor 2.1.15 update.
- Fabric
- Firmware updates on older disk servers.
Entries in GOC DB starting since the last report. |
Service | Scheduled? | Outage/At Risk | Start | End | Duration | Reason |
---|---|---|---|---|---|---|
lcgfts3.gridpp.rl.ac.uk, | SCHEDULED | WARNING | 30/11/2016 11:00 | 30/11/2016 13:00 | 2 hours | Upgrade of FTS3 service |
Open GGUS Tickets (Snapshot during morning of meeting) |
GGUS ID | Level | Urgency | State | Creation | Last Update | VO | Subject |
---|---|---|---|---|---|---|---|
125157 | Green | Less Urgent | Waiting for Reply | 2016-11-24 | 2016-11-29 | Creation of a repository within the EGI CVMFS infrastructure | |
124876 | Green | Less Urgent | On Hold | 2016-11-07 | 2016-11-21 | OPS | [Rod Dashboard] Issue detected : hr.srce.GridFTP-Transfer-ops@gridftp.echo.stfc.ac.uk |
124606 | Red | Top Priority | In Progress | 2016-10-24 | 2016-11-24 | CMS | Consistency Check for T1_UK_RAL |
124478 | Green | Less Urgent | In Progress | 2016-11-18 | 2016-11-18 | Jobs submitted via RAL WMS stuck in state READY forever and ever and ever | |
122827 | Green | Less Urgent | In Progress | 2016-07-12 | 2016-10-11 | SNO+ | Disk area at RAL |
117683 | Red | Less Urgent | On Hold | 2015-11-18 | 2016-10-05 | CASTOR at RAL not publishing GLUE 2 (Updated. There are ongoing discussions with GLUE & WLCG) |
Availability Report |
Key: Atlas HC = Atlas HammerCloud (Queue ANALY_RAL_SL6, Template 808); CMS HC = CMS HammerCloud
Day | OPS | Alice | Atlas | CMS | LHCb | Atlas HC | CMS HC | Comment |
---|---|---|---|---|---|---|---|---|
30/11/16 | 100 | 100 | 100 | 100 | 100 | N/A | 100 | |
01/12/16 | 100 | 100 | 100 | 99 | 100 | N/A | 100 | Single SRM test failure: User timeout over |
02/12/16 | 100 | 100 | 100 | 74 | 100 | N/A | 100 | Block of SRM test failures during the day (User timeout). |
03/12/16 | 100 | 100 | 100 | 99 | 100 | N/A | 100 | Single SRM test failure: User timeout over |
04/12/16 | 100 | 100 | 100 | 100 | 100 | N/A | 100 | |
05/12/16 | 100 | 100 | 100 | 98 | 100 | N/A | 100 | Single SRM test failure: User timeout over |
06/12/16 | 100 | 100 | 100 | 100 | 100 | N/A | 98 |
Notes from Meeting. |
- None Yet