Difference between revisions of "RALnonLHCCVMFS"

From GridPP Wiki
Jump to: navigation, search
(remove references to cvmfs v2.0)
Line 1: Line 1:
 
===CernmVM-FS Stratum 0 for non-LHC VOs ===
 
===CernmVM-FS Stratum 0 for non-LHC VOs ===
  
RAL Tier1 is currently hosting CernVM-FS Stratum-0 repositories for several non-LHC VOs.
+
RAL Tier-1 is currently hosting CernVM-FS Stratum-0 repositories for several non-LHC VOs.
  
 +
<!--
 
{| border="1"
 
{| border="1"
 
|+ Stratum-0 repos status at RAL
 
|+ Stratum-0 repos status at RAL
Line 46: Line 47:
 
| yes
 
| yes
 
|-|}
 
|-|}
 
+
-->
 
+
The Stratum-0 v2.1 repositories are replicated by the dedicated EGI CernVM-FS Stratum-1 server located at RAL.
Stratum-0 v2.0 repositories are replicated on the local LHC Stratum-1 server. Stratum-0 v2.1 repositories are replicated by the dedicated EGI CernVM-FS Stratum-1 server located at RAL.
+
  
 
==== Setting up a site to access small VOs CernVM-FS repositories located at RAL ====
 
==== Setting up a site to access small VOs CernVM-FS repositories located at RAL ====
  
Assuming that your site is already supporting CernVM-FS (client v2.1.X) for LHC VOs, these are the configurations needed to access the na62, mice, hone, wenmr, phys-ibergrid, hyperk, t2k, glast and cernatschool CernVM-FS areas.
+
Assuming that your site is already supporting CernVM-FS (client v2.1.X) for LHC VOs, these are the configurations needed to access the na62, mice, hone, wenmr, phys-ibergrid, hyperk, t2k, glast, cernatschool and biomed CernVM-FS areas.
  
 
On each worker node create the <code>/etc/cvmfs/domain.d/gridpp.ac.uk.conf</code> file
 
On each worker node create the <code>/etc/cvmfs/domain.d/gridpp.ac.uk.conf</code> file
Line 122: Line 122:
 
|}
 
|}
  
 +
<!--
 
<b>Note 2:</b> If your CernVM-FS clients are at v2.0.X, then on each WN, in <code>/etc/cvmfs/default.local</code> the <code>CVMFS_REPOSITORIES</code> variable should contain the names of the repositories i.e.
 
<b>Note 2:</b> If your CernVM-FS clients are at v2.0.X, then on each WN, in <code>/etc/cvmfs/default.local</code> the <code>CVMFS_REPOSITORIES</code> variable should contain the names of the repositories i.e.
  
Line 140: Line 141:
  
 
in <code>/etc/cvmfs/domain.d/gridpp.ac.uk.conf</code> file. As a consequence the glast, t2k, hyperk and cernatschool repositories will not be accessible by CernVM-FS clients v2.0 (upgrade to v2.1.X is recommended).
 
in <code>/etc/cvmfs/domain.d/gridpp.ac.uk.conf</code> file. As a consequence the glast, t2k, hyperk and cernatschool repositories will not be accessible by CernVM-FS clients v2.0 (upgrade to v2.1.X is recommended).
 
+
-->
  
 
In addition, also in <code>/etc/cvmfs/default.local</code> you must specify the names of your local site squids (regardless CernVM-FS client version):
 
In addition, also in <code>/etc/cvmfs/default.local</code> you must specify the names of your local site squids (regardless CernVM-FS client version):
Line 163: Line 164:
 
</pre>
 
</pre>
  
A RPM installing the public key and the gridpp.ac.uk.conf file is available and can be provided at request (catalin.condurache@stfc.ac.uk)
+
A RPM installing the public key and the gridpp.ac.uk.conf file is available and can be provided on request (catalin.condurache@stfc.ac.uk)
 
<!-- It will be made available shortly with a RPM package. -->
 
<!-- It will be made available shortly with a RPM package. -->
  

Revision as of 14:48, 25 April 2014

CernmVM-FS Stratum 0 for non-LHC VOs

RAL Tier-1 is currently hosting CernVM-FS Stratum-0 repositories for several non-LHC VOs.

The Stratum-0 v2.1 repositories are replicated by the dedicated EGI CernVM-FS Stratum-1 server located at RAL.

Setting up a site to access small VOs CernVM-FS repositories located at RAL

Assuming that your site is already supporting CernVM-FS (client v2.1.X) for LHC VOs, these are the configurations needed to access the na62, mice, hone, wenmr, phys-ibergrid, hyperk, t2k, glast, cernatschool and biomed CernVM-FS areas.

On each worker node create the /etc/cvmfs/domain.d/gridpp.ac.uk.conf file

[root@lcg0999 ~]# cat /etc/cvmfs/domain.d/gridpp.ac.uk.conf
CVMFS_PUBLIC_KEY=/etc/cvmfs/keys/gridpp.ac.uk.pub
CVMFS_SERVER_URL="http://cvmfs-egi.gridpp.rl.ac.uk:8000/cvmfs/@org@.gridpp.ac.uk;http://cvmfs01.nikhef.nl/cvmfs/@org@.gridpp.ac.uk"

Note 1: If you're using puppet module to configure the CernVM-FS clients, then it is not possible to define 'gridpp.ac.uk' CernVM-FS domain for the time being. Only solution is configuration per VO. The table below shows the current status at UK sites.

CernVM-FS clients configuration at UK sites
Site gridpp.ac.uk domain per VO configuration
RAL Tier-1 yes
RALPP yes
ICL yes
LIV
QMUL yes(?)
OX yes
GLA yes
SHEF
MANC yes


In addition, also in /etc/cvmfs/default.local you must specify the names of your local site squids (regardless CernVM-FS client version):

CVMFS_HTTP_PROXY="http://first.squid.domain:3128;http://second.squid.domain:3128"

We use a single pair of masterkeys (private and public) to sign all cvmfs stratum-0 repositories and the public key is needed for authentication at cvmfs client level.

[root@lcg1060 ~]# cat /etc/cvmfs/keys/gridpp.ac.uk.pub
-----BEGIN PUBLIC KEY-----
MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAp7C4KDvOIEVJepuAHjxE
EES1sDdohz0hiU6uvSqxVYjKVR4Y4/0I/D/zLijQI+MHR7859RN0/6fsZ3b3At3l
UbvNfqq6DN1zVjjd0xagC6SMBhSfj/iQKQSsG8MXSyiNmM8YalVHJSPqoova6CPE
EgLEjnHKTNEogTNjKBwbP2ELPLkfVoNoxxrXPSox7aln8JdgyZzZlBwm98gnFa1v
JTVAl0HQnUJ6cjMwO31wIGVMdvZ+P962t+2bPGfOCm6Ly6BusXcLoIIeez5SBerB
aHz//NSTZDbHVNPEqpoo1AQVVOo4XJmqo64jBa3G4Dr0zSda1bkZMVhsyUtjhfEB
DwIDAQAB
-----END PUBLIC KEY-----

A RPM installing the public key and the gridpp.ac.uk.conf file is available and can be provided on request (catalin.condurache@stfc.ac.uk)

The last requirement is that the site local squid(s) are configured to use the stratum-1 server at RAL.

Also do not forget to define the following environment variables:

- for mice users/jobs

VO_MICE_SW_DIR=/cvmfs/mice.gridpp.ac.uk

- for na62 users/jobs

VO_NA62_VO_GRIDPP_AC_UK_SW_DIR=/cvmfs/na62.gridpp.ac.uk

- for hone users/jobs

VO_HONE_SW_DIR=/cvmfs/hone.gridpp.ac.uk

- for enmr.eu users/jobs

VO_ENMR_EU_SW_DIR=/cvmfs/wenmr.gridpp.ac.uk

- for phys.vo.ibergrid.eu users/jobs

VO_PHYS_VO_IBERGRID_EU_SW_DIR=/cvmfs/phys-ibergrid.gridpp.ac.uk

- for hyperk.org users/jobs

VO_HYPERK_ORG_SW_DIR=/cvmfs/hyperk.gridpp.ac.uk

- for t2k.org users/jobs

VO_T2K_ORG_SW_DIR=/cvmfs/t2k.gridpp.ac.uk

- for cernatschool.org users/jobs

VO_CERNATSCHOOL_ORG_SW_DIR=/cvmfs/cernatschool.gridpp.ac.uk

- for glast.org users/jobs

VO_GLAST_ORG_SW_DIR=/cvmfs/glast.gridpp.ac.uk