|
|
Line 15: |
Line 15: |
| <!-- *********************************************************** -----> | | <!-- *********************************************************** -----> |
| | | |
− | ====== ======
| |
− | <!-- ******************************************************* ----->
| |
− | <!-- ***********Start Resolved Disk Server Issues*********** ----->
| |
− | {| width="100%" cellspacing="0" cellpadding="0" style="background-color: #ffffff; border: 1px solid silver; border-collapse: collapse; width: 100%; margin: 0 0 1em 0;"
| |
− | |-
| |
− | | style="background-color: #f8d6a9; border-bottom: 1px solid silver; text-align: center; font-size: 1em; font-weight: bold; margin-top: 0; margin-bottom: 0; padding-top: 0.1em; padding-bottom: 0.1em;" | Resolved Disk Server Issues
| |
− | |}
| |
− | * Newly deployed server GDSS721 (LHCbDst - D1T0) was taken back out of service following problems on 20th March. It was put back in production the following day after the problems were traced to errors in the network routing table and fixed.
| |
− | <!-- ***********End Resolved Disk Server Issues*********** ----->
| |
− | <!-- ***************************************************** ----->
| |
| | | |
− | ====== ======
| |
− | <!-- ***************************************************************** ----->
| |
− | <!-- ***********Start Current operational status and issues*********** ----->
| |
− | {| width="100%" cellspacing="0" cellpadding="0" style="background-color: #ffffff; border: 1px solid silver; border-collapse: collapse; width: 100%; margin: 0 0 1em 0;"
| |
− | |-
| |
− | | style="background-color: #b7f1ce; border-bottom: 1px solid silver; text-align: center; font-size: 1em; font-weight: bold; margin-top: 0; margin-bottom: 0; padding-top: 0.1em; padding-bottom: 0.1em;" | Current operational status and issues
| |
− | |}
| |
− | * There have been problems with the CMS Castor instance in recent weeks. These are triggered by high load. Work is underway to alleviate these problems, in particular servers with faster network connections will be moved into the disk cache in front of CMS_Tape when they become available.
| |
− | * The Castor Team are now able to reproduce the intermittent failures of Castor access via the SRM that has been reported in recent weeks. Understanding of the problem is significantly adcanced and further investigations are ongoing using the Castor Preprod instance. Ideas for a workaround are being developed.
| |
− | * As reported before, working with Atlas the file deletion rate was somewhat improved. However, there is still a problem that needs to be understood.
| |
− | * Around 50 files in tape backed service classes (mainly in GEN) have been found not to have migrated to tape. This is under investigation. The cause for some of these is understood (a bad tape at time of migration).
| |
− | * Problems with the infrastructure used to host many of our non-Catsor services have largely been worked around, although not yet fixed. Some additional migrations of VMs has been necessary.
| |
− | <!-- ***********End Current operational status and issues*********** ----->
| |
− | <!-- *************************************************************** ----->
| |
| | | |
| ====== ====== | | ====== ====== |
Key: Atlas HC = Atlas HammerCloud (Queue ANALY_RAL_SL6, Template 508); CMS HC = CMS HammerCloud