Difference between revisions of "RAL Tier1 weekly operations castor 09/02/2015"
From GridPP Wiki
(Created page with " [https://www.gridpp.ac.uk/wiki/RAL_Tier1_weekly_operations_castor List of CASTOR meetings] == Operations News == * Draining - Atlas currently on hold due to lack of free...") |
|||
Line 3: | Line 3: | ||
== Operations News == | == Operations News == | ||
− | + | * Draining - Atlas now draining again / LHCb also draining | |
− | + | * Tier 1 CASTOR stop and rebooted for Ghost vulnerability (and CIP) | |
− | + | * Facilities CASTOR rebooted for Ghost vulnerability | |
− | + | ||
− | * Draining - Atlas | + | |
− | * | + | |
− | * CASTOR | + | |
− | + | ||
− | + | ||
− | + | ||
== Operations Problems == | == Operations Problems == | ||
* There is a problem with latest gfal libraries - not a new problem but Atlas are starting to exercise functionality and identifying these issues | * There is a problem with latest gfal libraries - not a new problem but Atlas are starting to exercise functionality and identifying these issues | ||
− | + | * Atlas are putting files (sonar.test files) into un-routable paths - this looks like an issue with the space token used | |
− | + | * storageD retrieval from castor problems - investigation ongoing | |
− | + | * CMS heavy load & high job failure – hot spotting 3 servers, files spread to many servers which then also became loaded. CMS moved away from RAL for these files. Need to discuss this issue in more detail. | |
* castor functional test on lcgccvm02 causing problems - Gareth reviewing | * castor functional test on lcgccvm02 causing problems - Gareth reviewing | ||
− | |||
* 150k zero size files reported last week have almost all been dealt with, CMS files outstanding | * 150k zero size files reported last week have almost all been dealt with, CMS files outstanding | ||
* Files with no ns or xattr checksum value in castor are failing transfers from RAL to BNL using the BNL FTS3 server. | * Files with no ns or xattr checksum value in castor are failing transfers from RAL to BNL using the BNL FTS3 server. | ||
− | |||
− | |||
− | |||
− | |||
Line 34: | Line 22: | ||
== Planned, Scheduled and Cancelled Interventions == | == Planned, Scheduled and Cancelled Interventions == | ||
− | + | * Oracle upgrade (11.2.0.4) of preprod 9th Feb - will require a short outage | |
− | * Oracle | + | * Oracle PSU update - Tuesday Primary / Wednesday standby |
− | + | ||
− | + | ||
* Upgrade Oracle DB to version 11.2.0.4 (Late February?) | * Upgrade Oracle DB to version 11.2.0.4 (Late February?) | ||
* Upgrade CASTOR to version 2.1.14-14 OR 2.1.14-15 (February) | * Upgrade CASTOR to version 2.1.14-14 OR 2.1.14-15 (February) | ||
Line 60: | Line 46: | ||
* Chris/Rob to arrange a meeting to discuss CMS performance/xroot issues (is performance appropriate, if not plan to resolve) - inc. Shaun, Rob, Brian, Gareth | * Chris/Rob to arrange a meeting to discuss CMS performance/xroot issues (is performance appropriate, if not plan to resolve) - inc. Shaun, Rob, Brian, Gareth | ||
* Gareth to investigate providing checks for /etc/noquatto on production nodes & checks for fetch-crl | * Gareth to investigate providing checks for /etc/noquatto on production nodes & checks for fetch-crl | ||
− | * Matt to identify a suitable time to patch CASTOR facilities - DB team will synchronise patching of Juno with this outage | + | |
+ | * Matt to identify a suitable time to patch CASTOR facilities - DB team will synchronise patching of Juno with this outage - Done | ||
== Staffing == | == Staffing == | ||
* Castor on Call person | * Castor on Call person | ||
− | ** Rob | + | ** Rob |
* Staff absence/out of the office: | * Staff absence/out of the office: | ||
− | + | ** Chris working from home Monday | |
− | ** Chris | + | |
− | + |
Latest revision as of 16:58, 6 February 2015
Contents
Operations News
- Draining - Atlas now draining again / LHCb also draining
- Tier 1 CASTOR stop and rebooted for Ghost vulnerability (and CIP)
- Facilities CASTOR rebooted for Ghost vulnerability
Operations Problems
- There is a problem with latest gfal libraries - not a new problem but Atlas are starting to exercise functionality and identifying these issues
- Atlas are putting files (sonar.test files) into un-routable paths - this looks like an issue with the space token used
- storageD retrieval from castor problems - investigation ongoing
- CMS heavy load & high job failure – hot spotting 3 servers, files spread to many servers which then also became loaded. CMS moved away from RAL for these files. Need to discuss this issue in more detail.
- castor functional test on lcgccvm02 causing problems - Gareth reviewing
- 150k zero size files reported last week have almost all been dealt with, CMS files outstanding
- Files with no ns or xattr checksum value in castor are failing transfers from RAL to BNL using the BNL FTS3 server.
Blocking Issues
- grid ftp bug in SL6 - stops any globus copy if a client is using a particular library. This is a show stopper for SL6 on disk server.
Planned, Scheduled and Cancelled Interventions
- Oracle upgrade (11.2.0.4) of preprod 9th Feb - will require a short outage
- Oracle PSU update - Tuesday Primary / Wednesday standby
- Upgrade Oracle DB to version 11.2.0.4 (Late February?)
- Upgrade CASTOR to version 2.1.14-14 OR 2.1.14-15 (February)
Advanced Planning
Tasks
- DB team need to plan some work which will result in the DBs being under load for approx 1h - not terribly urgent but needs to be done in new year.
- Provide new VM? to provide castor client functionality to query the backup DBs
- Plan to ensure PreProd represents production in terms of hardware generation are underway
- Possible future upgrade to CASTOR 2.1.14-15 post-Christmas
- Switch from admin machines: lcgccvm02 to lcgcadm05
- Correct partitioning alignment issue (3rd CASTOR partition) on new castor disk servers
Interventions
Actions
- Rob to pick up DB cleanup change control
- Bruno to document processes to control services previously controlled by puppet
- Gareth to arrange meeting castor/fab/production to discuss the decommissioning procedures
- Chris/Rob to arrange a meeting to discuss CMS performance/xroot issues (is performance appropriate, if not plan to resolve) - inc. Shaun, Rob, Brian, Gareth
- Gareth to investigate providing checks for /etc/noquatto on production nodes & checks for fetch-crl
- Matt to identify a suitable time to patch CASTOR facilities - DB team will synchronise patching of Juno with this outage - Done
Staffing
- Castor on Call person
- Rob
- Staff absence/out of the office:
- Chris working from home Monday