<!-- keep this as a security measure: * Set ALLOWTOPICCHANGE = Main.TWikiAdminGroup,Main.LCGAdminGroup,Main.EgiGroup * Set ALLOWTOPICRENAME = Main.TWikiAdminGroup,Main.LCGAdminGroup #uncomment this if you want the page only be viewable by the internal people #* Set ALLOWTOPICVIEW = Main.TWikiAdminGroup,Main.LCGAdminGroup --> ---+ Swiss Grid Operations Meeting on 2013-04-04 * *Date and time*: First Thursday of the month, at 14:00 * *Place*: Vidyo (room: Swiss_Grid_Operations_Meeting, extension: 9227296) * *External link*: http://vidyoportal.cern.ch/flex.html?roomdirect.html&key=Nrq24qRR4V1u * *Phone gate*: From Switzerland: 0225330322 (portal) + 9227296 (extension) + # (pound sign) * *IRC chat*: irc:gridchat.cscs.ch:994#lcg (ask pw via email) ---++ Agenda Status * CSCS (reports Pablo): * F2F meeting held @ CSCS: MeetingCHIPPCSCSFaceToFace20130321 * SRM security issue striked again, still no answer from dCache support. Upgrade in one month. * CMS Argus authentication issues still under investigation * Cream03 with EMI-2 and SL6 in production. Feedback? * cms01 ready * Preparation for Maintenance on Tuesday next week: SiteMaintenance20130409 * PSI (reports Fabio): * SL6 kernels upgraded to =2.6.32-358.2.1= but 2 files servers are affected by a load reporting issue [[https://bugzilla.redhat.com/show_bug.cgi?id=883905][xfsaild always in D-state]]. * [[http://support.netapp.com/NOW/public/apbu/oemcp/][RDAC]] driver upgraded from =09.03.0C08.0535= to =09.03.0C05.0642= on these 2 files servers. * SL5 kernels upgraded to =2.6.18-348.3.1=. * Installed by Yum the [[http://xrootd.slac.stanford.edu/dload.html][xrootd clients]] at version =3.3.1-1= on each UI and WN. * Postgresql =9.2.2= upgraded to, at that time, latest =9.2.3-2= while now the latest is =9.2.4=. * *dCache migrated* from =1.9.12-23= to =2.2.9-1=: major news ( look PSI files reported below ): * Postgresl max # connections raised from =200= to =300=, you can use this Nagios check to be alerted about them [[http://bucardo.org/wiki/Check_postgres][/usr/bin/check_postgres.pl --action=backends]] * Replaced =ssh1= admin door with the =ssh2= door. * Adapted =dCache_gmetric.py= to manage the new =ssh2= case. * Replaced gPlazma1 with [[http://www.dcache.org/manuals/2012/workshop/slides/gPlazma2_configuration.pdf][gPlazma2]]: migration was not transparent and I got a [[https://ggus.eu/ws/ticket_info.php?ticket=93009][GGUS Ticket]], now solved. * New Pool Selection algorithm = [[https://indico.desy.de/getFile.py/access?contribId=7&resId=0&materialId=slides&confId=5289][Weighted Available Space Selection]]. Needs two changes, the wass and the weights. * The Derek's tools are still compatible. * News about BDII, look =/etc/dcache/dcache.conf=. Does not work on the infoprovider layout. * *When you start dCache you have to wait 10s and then*: =/etc/init.d/dcache-server restart t3se01-Domain-utility=. It's a bug they are working out. * I can't make work file =/var/log/dcache/adminshell_history= * I can't make work gPlazma2 plugin =nsswitch=. * Need to test Secondary groups with =gPlazma2=, it fails with =gPlazma1=. * Raised to the =info= level each cell, nice to follow what was happening before a crash/error, look the =logback.xml= file. * Enabled [[https://twiki.cern.ch/twiki/bin/view/Main/CmsXrootdArchitecture][Xrootd]] at PSI, I can copy files and I dedicated an I/O queue in dCache but I'm still missing all the details, so it's not at the production level now. It's also not strictly required for a T3 so it has low priority. * UNIBE (reports Gianfranco - won't be able to attend): * 5 out of 6 DPM disk servers are now SLC6.3, emi-dpm_disk 1.8.6-1.el6 (EMI-2) * New disk server with 120TB delivered, running acceptance tests, will deploy on Monday * We'll have 502TB (after RAID6): 400TB ATLASDATADISK (350TB pledged), 80TB LOCALGROUPDISK, 22TB for ops, SMSCG VOs * SLC5 and SLC6 kernel upgrade on interactive nodes/grid UI. * Cluster not upgraded, don't want to risk yet another ARC bug (there is already one related to the new kernel, which should not affect us). Risk very low: * all WNs in a private network * no interactive submission allowed * no users except grid pool accounts * only atlas/ch, smscg VOs, Andrej Filipcic allowed via arc.conf and grid-mapfile Next meeting date: 2nd of May ---++ Attendants * CSCS: Pablo * CMS: Fabio, Daniel, Derek * ATLAS: * LHCb: Roland * EGI: ---++ Action items * Item1 ---++ PSI dCache 2.2 files ---+++ /etc/dcache/logback.xml %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre> /etc/dcache/logback.xml:177: <appender>stdout</appender> /etc/dcache/logback.xml-178- <logger>root</logger> /etc/dcache/logback.xml-179- <level>info</level> <---------- </pre> %ENDTWISTY% ---+++ /var/lib/dcache/config/poolmanager.conf %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># # Created by PoolManager(PoolManager) at Fri Mar 01 16:05:25 CET 2013 # # # Submodule CostModule (cm) : diskCacheV111.poolManager.CostModuleV1 # $Revision: 16335 $ # cm set debug off cm set update on cm set magic on # note new wass type pm create -type=wass default pm set default -cpucostfactor=1.0 -spacecostfactor=1.0 pm set default -idle=0.0 -p2p=0.4 -alert=0.0 -halt=0.0 -fallback=0.0 pm set default -p2p-allowed=yes -p2p-oncost=yes -p2p-fortransfer=yes pm set default -stage-allowed=no pm set default -max-copies=500 -slope=0.0 # # Setup of PoolManager (diskCacheV111.poolManager.PoolManagerV5) at Fri Mar 01 16:05:25 CET 2013 # # # Printed by diskCacheV111.poolManager.PoolSelectionUnitV2 at Fri Mar 01 16:05:25 CET 2013 # # psu set regex off psu set allpoolsactive off # # The units ... # psu create unit -protocol */* psu create unit -store dteam:dteam@osm psu create unit -store ops:ops@osm psu create unit -net 0.0.0.0/0.0.0.0 psu create unit -store cms:cms@osm psu create unit -store *@* # # The unit Groups ... # psu create ugroup any-store psu addto ugroup any-store *@* psu create ugroup cms_unit_group psu addto ugroup cms_unit_group cms:cms@osm psu create ugroup dteam_unit_group psu addto ugroup dteam_unit_group dteam:dteam@osm psu create ugroup world-net psu addto ugroup world-net 0.0.0.0/0.0.0.0 psu create ugroup any-protocol psu addto ugroup any-protocol */* psu create ugroup ops_unit_group psu addto ugroup ops_unit_group ops:ops@osm # # The pools ... # CMS POOLs #psu create pool t3se01_cms #psu create pool t3se01_cms_1 #psu create pool t3se01_ops psu create pool t3fs01_cms psu create pool t3fs02_cms psu create pool t3fs03_cms psu create pool t3fs04_cms psu create pool t3fs04_cms_1 psu create pool t3fs07_cms psu create pool t3fs07_cms_1 psu create pool t3fs08_cms psu create pool t3fs08_cms_1 psu create pool t3fs09_cms psu create pool t3fs09_cms_1 psu create pool t3fs10_cms psu create pool t3fs10_cms_1 psu create pool t3fs11_cms psu create pool t3fs11_cms_1 psu create pool t3fs13_cms psu create pool t3fs13_cms_1 psu create pool t3fs13_cms_2 psu create pool t3fs13_cms_3 psu create pool t3fs13_cms_4 psu create pool t3fs13_cms_5 psu create pool t3fs13_cms_6 psu create pool t3fs14_cms psu create pool t3fs14_cms_1 psu create pool t3fs14_cms_2 psu create pool t3fs14_cms_3 psu create pool t3fs14_cms_4 psu create pool t3fs14_cms_5 psu create pool t3fs14_cms_6 # OPS POOLs psu create pool t3fs01_ops psu create pool t3fs02_ops psu create pool t3fs03_ops psu create pool t3fs04_ops psu create pool t3fs07_ops psu create pool t3fs08_ops psu create pool t3fs09_ops psu create pool t3fs10_ops psu create pool t3fs11_ops psu create pool t3fs13_ops psu create pool t3fs13_ops_1 psu create pool t3fs13_ops_2 psu create pool t3fs13_ops_3 psu create pool t3fs13_ops_4 psu create pool t3fs13_ops_5 psu create pool t3fs14_ops psu create pool t3fs14_ops_1 psu create pool t3fs14_ops_2 psu create pool t3fs14_ops_3 psu create pool t3fs14_ops_4 psu create pool t3fs14_ops_5 # # The pool groups ... # psu create pgroup dteam #psu addto pgroup dteam t3se01_ops psu addto pgroup dteam t3fs01_ops psu addto pgroup dteam t3fs02_ops psu addto pgroup dteam t3fs03_ops psu addto pgroup dteam t3fs04_ops psu addto pgroup dteam t3fs07_ops psu addto pgroup dteam t3fs08_ops psu addto pgroup dteam t3fs09_ops psu addto pgroup dteam t3fs10_ops psu addto pgroup dteam t3fs11_ops psu addto pgroup dteam t3fs13_ops psu addto pgroup dteam t3fs13_ops_1 psu addto pgroup dteam t3fs13_ops_2 psu addto pgroup dteam t3fs13_ops_3 psu addto pgroup dteam t3fs13_ops_4 psu addto pgroup dteam t3fs13_ops_5 psu addto pgroup dteam t3fs14_ops psu addto pgroup dteam t3fs14_ops_1 psu addto pgroup dteam t3fs14_ops_2 psu addto pgroup dteam t3fs14_ops_3 psu addto pgroup dteam t3fs14_ops_4 psu addto pgroup dteam t3fs14_ops_5 # psu create pgroup cms #psu addto pgroup cms t3se01_cms #psu addto pgroup cms t3se01_cms_1 psu addto pgroup cms t3fs01_cms psu addto pgroup cms t3fs02_cms psu addto pgroup cms t3fs03_cms psu addto pgroup cms t3fs04_cms psu addto pgroup cms t3fs04_cms_1 psu addto pgroup cms t3fs07_cms psu addto pgroup cms t3fs07_cms_1 psu addto pgroup cms t3fs08_cms psu addto pgroup cms t3fs08_cms_1 psu addto pgroup cms t3fs09_cms psu addto pgroup cms t3fs09_cms_1 psu addto pgroup cms t3fs10_cms psu addto pgroup cms t3fs10_cms_1 psu addto pgroup cms t3fs11_cms psu addto pgroup cms t3fs11_cms_1 psu addto pgroup cms t3fs13_cms psu addto pgroup cms t3fs13_cms_1 psu addto pgroup cms t3fs13_cms_2 psu addto pgroup cms t3fs13_cms_3 psu addto pgroup cms t3fs13_cms_4 psu addto pgroup cms t3fs13_cms_5 psu addto pgroup cms t3fs13_cms_6 psu addto pgroup cms t3fs14_cms psu addto pgroup cms t3fs14_cms_1 psu addto pgroup cms t3fs14_cms_2 psu addto pgroup cms t3fs14_cms_3 psu addto pgroup cms t3fs14_cms_4 psu addto pgroup cms t3fs14_cms_5 psu addto pgroup cms t3fs14_cms_6 psu create pgroup ops #psu addto pgroup ops t3se01_ops psu addto pgroup ops t3fs01_ops psu addto pgroup ops t3fs02_ops psu addto pgroup ops t3fs03_ops psu addto pgroup ops t3fs04_ops psu addto pgroup ops t3fs07_ops psu addto pgroup ops t3fs08_ops psu addto pgroup ops t3fs09_ops psu addto pgroup ops t3fs10_ops psu addto pgroup ops t3fs11_ops psu addto pgroup ops t3fs13_ops psu addto pgroup ops t3fs13_ops_1 psu addto pgroup ops t3fs13_ops_2 psu addto pgroup ops t3fs13_ops_3 psu addto pgroup ops t3fs13_ops_4 psu addto pgroup ops t3fs13_ops_5 psu addto pgroup ops t3fs14_ops psu addto pgroup ops t3fs14_ops_1 psu addto pgroup ops t3fs14_ops_2 psu addto pgroup ops t3fs14_ops_3 psu addto pgroup ops t3fs14_ops_4 psu addto pgroup ops t3fs14_ops_5 # # The links ... # psu create link cms-link cms_unit_group world-net psu set link cms-link -readpref=10 -writepref=10 -cachepref=0 -p2ppref=10 psu add link cms-link cms psu create link ops-link ops_unit_group world-net psu set link ops-link -readpref=10 -writepref=10 -cachepref=0 -p2ppref=10 psu add link ops-link ops psu create link dteam-link dteam_unit_group world-net psu set link dteam-link -readpref=10 -writepref=10 -cachepref=0 -p2ppref=10 psu add link dteam-link dteam # # The link Groups ... # psu create linkGroup cms-linkGroup psu set linkGroup custodialAllowed cms-linkGroup false psu set linkGroup replicaAllowed cms-linkGroup true psu set linkGroup nearlineAllowed cms-linkGroup false psu set linkGroup outputAllowed cms-linkGroup true psu set linkGroup onlineAllowed cms-linkGroup true psu addto linkGroup cms-linkGroup cms-link psu create linkGroup dteam-linkGroup psu set linkGroup custodialAllowed dteam-linkGroup false psu set linkGroup replicaAllowed dteam-linkGroup true psu set linkGroup nearlineAllowed dteam-linkGroup false psu set linkGroup outputAllowed dteam-linkGroup true psu set linkGroup onlineAllowed dteam-linkGroup true psu addto linkGroup dteam-linkGroup dteam-link psu create linkGroup ops-linkGroup psu set linkGroup custodialAllowed ops-linkGroup false psu set linkGroup replicaAllowed ops-linkGroup true psu set linkGroup nearlineAllowed ops-linkGroup false psu set linkGroup outputAllowed ops-linkGroup true psu set linkGroup onlineAllowed ops-linkGroup true psu addto linkGroup ops-linkGroup ops-link # # Submodule [rc] : class diskCacheV111.poolManager.RequestContainerV5 # rc onerror suspend rc set max retries 3 rc set retry 900 rc set warning path billing rc set poolpingtimer 600 rc set max restore unlimited rc set sameHostCopy besteffort rc set sameHostRetry notchecked rc set max threads 2147483647 # </pre> %ENDTWISTY% ---+++ /etc/dcache/LinkGroupAuthorization.conf ( note the * ) %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># Puppet Managed File LinkGroup cms-linkGroup /cms LinkGroup ops-linkGroup /ops/NGI/Germany* /ops/NGI/Switzerland* LinkGroup dteam-linkGroup /dteam /dteam/Role=NULL/Capability=NULL </pre> %ENDTWISTY% ---+++ /etc/dcache/dcache.conf %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># cat /etc/dcache/dcache.conf # Puppet Managed File dcache.layout=${host.name} dcache.namespace=chimera #chimera.db.user = postgres #chimera.db.url = jdbc:postgresql://localhost/chimera?prepareThreshold=3 chimera.db.user = chimera chimera.db.url = jdbc:postgresql://t3dcachedb03.psi.ch/chimera?prepareThreshold=3 # The following is taken from the old dCacheSetup file. # Some configuration parameters may no longer apply. # Dedicated user for dcache, not anymore root dcache.user=dcache dcache.paths.billing=/var/log/dcache # To check permissions not just in the dir where we are but also in the upper dirs pnfsVerifyAllLookups=true dcache.java.memory.heap=1024m dcache.java.memory.direct=1024m #pool.dcap.port=0 net.inetaddr.lifetime=1800 net.wan.port.min=20000 net.wan.port.max=25000 net.lan.port.min=33115 net.lan.port.max=33145 broker.host=t3se01.psi.ch ###### POOL VARs ###### # poolIoQueue is defined in share/defaults/pool.properties and used by share/services/pool.batch ; the queue 'regular' is always created by default poolIoQueue=wan,xrootd waitForFiles=${path}/setup lfs=precious tags=hostname=${host.name} ####################### metaDataRepository=org.dcache.pool.repository.meta.db.BerkeleyDBMetaDataRepository useGPlazmaAuthorizationModule=false useGPlazmaAuthorizationCell=true # gsidcapIoQueue is defined in share/defaults/dcap.properties and used by share/services/gsidcap.batch #gsidcapIoQueue=default # dcapIoQueue is defined in share/defaults/dcap.properties and used by share/services/dcap.batch #dcapIoQueue=default ##performanceMarkerPeriod=10 # gsiftpIoQueue is used by share/services/gridftp.batch gsiftpIoQueue=wan xrootdIoQueue=xrootd ###### SRM VARs ###### # remoteGsiftpIoQueue is defined in share/defaults/srm.properties and used by share/services/srm.batch remoteGsiftpIoQueue=wan srmDatabaseHost=t3dcachedb03.psi.ch srmDbName=dcache srmDbUser=srmdcache srmDbPassword= srmSpaceManagerEnabled=yes # ---- Log to database # # # If set to true, the transfer services log transfers to the srm # database. srmDbLogEnabled=true # ---- Enables SRM request transition history logging # # Enables logging of transition history of SRM request in the # database. The request transitions can be examined through the # command line interface or through the the srmWatch monitoring tool. # # Enabling this feature increases the size and load of the database. srmRequestHistoryDatabaseEnabled=true ###################### ftpPort=${portBase}126 kerberosFtpPort=${portBase}127 companionDatabaseHost=t3dcachedb03.psi.ch spaceManagerDatabaseHost=t3dcachedb03.psi.ch pinManagerDbHost=t3dcachedb03.psi.ch defaultPnfsServer=t3dcachedb03.psi.ch SpaceManagerReserveSpaceForNonSRMTransfers=true SpaceManagerLinkGroupAuthorizationFileName=/etc/dcache/LinkGroupAuthorization.conf dcache.log.dir=/var/log/dcache billingToDb=yes billingDbHost=t3dcachedb03.psi.ch billingDbUser=srmdcache billingDbPass= billingDbName=billing billingMaxInsertsBeforeCommit=10000 billingMaxTimeBeforeCommitInSecs=5 # info service properties info-provider.site-unique-id=T3_CH_PSI info-provider.se-unique-id=t3se01.psi.ch info-provider.se-name=SRM endpoint for T3_CH_PSI info-provider.glue-se-status=Production info-provider.dcache-quality-level=production info-provider.dcache-architecture=multidisk </pre> %ENDTWISTY% ---+++ /etc/dcache/layouts/t3se.conf %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># Puppet Managed File [${host.name}-Domain-dcap] [${host.name}-Domain-dcap/dcap] [${host.name}-Domain-gridftp] [${host.name}-Domain-gridftp/gridftp] [${host.name}-Domain-gsidcap] [${host.name}-Domain-gsidcap/gsidcap] [${host.name}-Domain-srm] [${host.name}-Domain-srm/srm] [${host.name}-Domain-srm/spacemanager] [${host.name}-Domain-srm/transfermanagers] [${host.name}-Domain-httpd] [${host.name}-Domain-httpd/httpd] [${host.name}-Domain-httpd/statistics] [${host.name}-Domain-httpd/billing] [${host.name}-Domain-httpd/srm-loginbroker] [${host.name}-Domain-utility] [${host.name}-Domain-utility/gsi-pam] [${host.name}-Domain-utility/pinmanager] [${host.name}-Domain-dir] [${host.name}-Domain-dir/dir] [${host.name}-Domain-info] [${host.name}-Domain-info/info] # to use WebDav you need to run on the Linux client # yum install --enablerepo=dag,epel davfs2 # mount.davfs http://t3se03:2880/pnfs/psi.ch/cms/ /pnfs/psi.ch/cms # #[${host.name}-Domain-webdav] #[${host.name}-Domain-webdav/webdav] #webdav.redirect.on-read=false #webdav.redirect.on-write=false #webdavRootPath=/ #webdavAllowedPaths=/pnfs/psi.ch/cms #webdavAnonymousAccess=READONLY #webdavReadOnly=true #webdavIoQueue=webdav #webdavProtocol=http [dCacheDomain] [dCacheDomain/poolmanager] poolmanager.cache-hit-messages.enabled=true [dCacheDomain/broadcast] [dCacheDomain/loginbroker] [dCacheDomain/topo] [${host.name}-Domain-xrootd] [${host.name}-Domain-xrootd/xrootd] useGPlazmaAuthorizationCell=false useGPlazmaAuthorizationModule=true poolmanager=${spacemanager} xrootdAuthNPlugin=gsi xrootdAllowedWritePaths= xrootdMoverTimeout=28800000 # Unauthenticated # xrootdPlugins=gplazma:none,authz:cms-tfc # Authenticated according to gplazma xrootdPlugins=gplazma:gsi,authz:cms-tfc # # Change this according to your location: xrootd.cms.tfc.path=/etc/xrootd/storage.xml # # Must be coherent with your TFC in storage.xml: xrootd.cms.tfc.protocol=root [${host.name}-Domain-CMS] [${host.name}-Domain-CMS/xrootd] loginBroker=srm-LoginBroker xrootdRootPath=/pnfs/psi.ch/cms/trivcat xrootdPort=1096 xrootdMoverTimeout=28800000 </pre> %ENDTWISTY% ---+++ /etc/dcache/layouts/t3dcachedb.conf %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># Puppet Managed File [${host.name}-Domain-gPlazma] [${host.name}-Domain-gPlazma/gplazma] gplazma.vorolemap.file=/etc/grid-security/grid-vorolemap [${host.name}-Domain-namespace] [${host.name}-Domain-namespace/pnfsmanager] [${host.name}-Domain-namespace/cleaner] [${host.name}-Domain-namespace/acl] [${host.name}-Domain-adminDoor] [${host.name}-Domain-adminDoor/admin] sshVersion=ssh2 admin.ssh2AdminPort=22224 adminHistoryFile=/var/log/dcache/adminshell_history [${host.name}-Domain-nfs] dcache.user=root [${host.name}-Domain-nfs/nfsv3] ## webadmin needs info Domain #[${host.name}-Domain-webadmin] #[${host.name}-Domain-webadmin/webadmin] #webadminHttpsPort=8082 #webadminHttpPort=8081 #webadminDCacheInstanceName=${host.name} ##webadminAuthenticated=true #webadminAdminGid=1000 #webadminAuthenticated=false </pre> %ENDTWISTY% ---+++ gPlazma2 /etc/dcache/gplazma.conf %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre>auth optional x509 auth optional voms map requisite vorolemap map requisite authzdb session requisite authzdb </pre> %ENDTWISTY% ---+++ vomsdir needed by gPlazma2 %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre># find /etc/grid-security/vomsdir/ /etc/grid-security/vomsdir/ /etc/grid-security/vomsdir/ops -> symbolic link to cms /etc/grid-security/vomsdir/cms /etc/grid-security/vomsdir/cms/voms.fnal.gov.lsc /etc/grid-security/vomsdir/cms/voms.cern.ch.lsc /etc/grid-security/vomsdir/cms/lcg-voms.cern.ch.lsc /etc/grid-security/vomsdir/dteam /etc/grid-security/vomsdir/dteam/voms2.hellasgrid.gr.lsc /etc/grid-security/vomsdir/dteam/voms.hellasgrid.gr.lsc </pre> %ENDTWISTY% ---+++ /etc/xrootd/xrootd-clustered.cfg %TWISTY{showlink="Show" hidelink="Hide" showimgleft="%ICONURLPATH{toggleopen-small}%" hideimgleft="%ICONURLPATH{toggleclose-small}%"}% <pre> # grep 64 /etc/xrootd/xrootd-clustered.cfg oss.namelib /usr/lib64/libXrdCmsTfc.so file:/etc/xrootd/storage.xml?protocol=direct xrootd.seclib /usr/lib64/libXrdSec.so xrootd.fslib /usr/lib64/libXrdOfs.so </pre> %ENDTWISTY%
This topic: LCGTier2
>
WebHome
>
MeetingsBoard
>
MeetingSwissGridOperations20130404
Topic revision: r13 - 2013-04-10 - FabioMartinelli
Copyright © 2008-2024 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki?
Send feedback