ATLAS


This is a basic page of documentation for ATLAS NAF users. The main NAF documentation can be found on this confluence page


Accounts


Getting an account


The NAF working environment is fully integrated into the standard DESY computing infrastructure: there is no distinction between a "DESY account" and a "NAF account".

There is however a dedicated unix group for ATLAS on the NAF, AF-ATLAS, which is usually added to all new ATLAS accounts.


The ATLAS group has a set of working group servers which are used as the NAF entry point, which can be accessed by any member of the DESY group AF-ATLAS.

If you are an existing DESY user with a full ATLAS account, you need to be added to the AF-ATLAS group in order to access NAF 2: send an email to naf-atlas-support@desy.de


If you previously had a DESY account but this has now been deactivated, including with another group, then rather than getting a new account the old one should be reactivated.

Please send an email to naf-atlas-support@desy.de with your first, last, and user names and specify that you would like to reactivate your account in order to get access to the NAF.


If you a non-DESY user, you can request a NAF account here.

NAF registration by non-DESY users requires a Grid certificate imported into your browser for authentication purposes.


More information on NAF accounts is available here, including how to reset passwords (required every 6 months).


Login


To login to the NAF you must be a member of the group AF-ATLAS. For login, simply do an ssh to the ATLAS NAF working group servers:

ssh -X -l yourusername naf-atlas.desy.de

In reality this is a load-balancing alias to a pool of eight working group servers, naf-atlas11-18, all running EL7.


Up to date information on the NAF working group servers and login is here.


Login to DESY and/or eduroam networks at DESY

You can login to the DESY or eduroam WiFi networks at DESY if your laptop is registered in QIP, and you have the correct resources on your account, namely "wlan_desy" and "eduroam" - send an email to naf-atlas-support@desy.de.


The login is via

<username>@desy.de

using a different password to your account password. Once you have the relevant groups, you can set this yourself at password.desy.de.


Finally, you also need to download the relevant "profiles" provided in the "Configuration of the WiFi Networks" section here.


NAF remote desktop login


You can use this browser based login to start a shell within the DESY firewall, for example to access internal DESY pages:


Point your browser at https://naf-x-el.desy.de:3389 (currently, an CentOS 7 system) or https://naf-x-ubuntu.desy.de:3389 (currently, an Ubuntu 20.04 system)


More details can be found here.


AFS quota


The afs cell used in the NAF is the standard DESY afs cell /afs/desy.de.

To learn about your current quota usage, do

fs lq

This can be done either in your afs $HOME or in $HOME/xxl (see here).


Account expiry


Accounts expire typically 3 years after creation, but may be extended if the owner continues to work on ATLAS with the same or a different ATLAS-D group.

Expiration notification e-mails are regularly sent to the DESY email address, beginning at 30 days before the expiration date.


If an extension is needed, please write to the namespace supervisors listed in the expiration notification e-mail, stating the username, the reason(s) for extension, and the additional length of time required.


Grid certificates


Information on how to get a DESY grid certificate from GÉANT is available here. Alternatively, a CERN grid certificate can be obtained here.


Deploying the certificate


To use the grid and corresponding software, make sure you have your usercert.pem and userkey.pem in your $HOME/.globus/ directory at DESY and/or CERN and that they have the right permissions.

A script is available here to extract the .pem files from your grid.certificate.p12. Copy the script, set the correct permissions for an executable using chmod 751, and then do:

CertConvert.sh -pem grid.certificate.p12

Many ATLAS interfaces such as Rucio, AMI and COMA use authentication via grid certificate and it is necessary to upload the certificate to your browser.

It is recommended to use the Firefox web-browser, and old certificates should be removed to avoid possible conflicts.


ATLAS VO membership


You will need to be a member of the ATLAS VO, membership can be applied for here. This will allow you to access ATLAS data and get a proxy.


Setting up a proxy


To use the grid certificate, you have to set up the grid environment first:

lsetup emi

This is also executed during Rucio setup. Now, you can create a proxy for the ATLAS-D role:

voms-proxy-init -voms atlas:/atlas/de

or get a proxy with the plain ATLAS role, use:

voms-proxy-init -voms atlas

To get information about your current proxy, for instance the time for how long it will stay valid:

voms-proxy-info -all


Storage on the NAF


DUST


The main local storage for NAF is DUST. Please note that this is "scratch" space, i.e. there is no back up.

DUST space is mounted on all ATLAS NAF working group servers and is visible from the batch system.


An inital 1 TB of space can be created for each NAF user under:

/nfs/dust/atlas/user/yourusername

Initial requests for DUST space and requests to increase the quota of existing DUST space should be made to naf-atlas-support@desy.de.

Please specify your username and in the case of an increase request how much additional space you require.


Group spaces on DUST

There are also several group spaces in the following directory:

/nfs/dust/atlas/group

where membership of a particular group can be applied for via naf-atlas-support@desy.de.

This is usually needed when people are working on the same analysis area and want to share files, although it should be noted that by default all files on DUST are readable by all users.


Please note: the group directories on DUST are set up to have read and write access for all members to all newly created sub-directories. So as a member, you can all create, read and delete directories and files in this group directory.


The current groups in the ATLAS DUST space are:

Group

Directory

Responsible

Description

AF-ATLZEED

/nfs/dust/atlas/group/zeed

Sasha Glazov

Z to ee analyses

AF-ATLZPRIME

/nfs/dust/atlas/group/zprime

Eleanor Jones

zprime analyses

AF-ATLJETMET

/nfs/dust/atlas/group/jetmet

Pavel Starovoitov

Jet-MET analyses

AF-ATLTOP

/nfs/dust/atlas/group/top

Timothée Theveneaux-Pelzer

Top analyses

AF-ATLEXTH

/nfs/dust/atlas/group/extendedhiggs

Claudia Seitz

Extended Higgs analysis

AF-ATLTTPH

/nfs/dust/atlas/group/ttbarphoton

Carmen Diez

ttbar plus photon analysis

AF-ATLWAI

/nfs/dust/atlas/group/wai

Ludovica Aperio Bella

W-Ai analysis

AF-ATLZAI

/nfs/dust/atlas/group/zai

Ludovica Aperio Bella

Z-Ai analysis

AF-ATLHTT

/nfs/dust/atlas/group/htautau

Karsten Koeneke

Higgs to tau tau analysis

AF-ATLHVNU

/nfs/dust/atlas/group/heavynu

Krisztian Peters

Heavy nu analysis

AF-ATLHEAVYH

/nfs/dust/atlas/group/ahtottbar

Katharina Behr

Heavy Higgs searches

AF-ATLTYC

/nfs/dust/atlas/group/topyukawa

Thorsten Kuhl

Top yukawa coupling

AF-TTENTANGLE

/nfs/dust/atlas/group/ttEntangle

Eleanor Jones

ttEntangle analysis


You can view your DUST usage and quota here from inside the DESY firewall, as well as on the command line via:

my-dust-quota -g

where the "-g" option also then displays the usage and quota of any groups to which you belong.


Take note of the DUST lifecycle policy detailed here, which governs how and when data on DUST are automatically removed.


Additional AFS space: xxl


An additional, larger afs space can be created for each NAF user, in a directory $HOME/xxl.

An initial 10 GB quota is applied, which may be increased up to a maximum of 30 GB, if necessary.


To get a NAF afs $HOME/xxl directory, please contact naf-helpdesk@desy.de


Grid storage: DESY-HH_LOCALGROUPDISK


If you run on the grid, the files of your output dataset are usually placed at the SCRATCHDISK of the site you where running (see here for information on grid certificates).


A large dCache based storage is provided for the NAF as a dedicated resource, DESY-HH_LOCALGROUPDISK, which visible from the NAF batch system and should be used to store large files.

An initial quota of 10TB is automatically assigned to all users with a german grid certificate via the /atlas/de role (see here). You can also request the /atlas/de role here.


Your current quota and usage on all Rucio Storage Elements (RSEs) can be checked here, and you can select DESY-HH_LOCALGROUPDISK (or any other) using the Search box.

Requests for additional quota on DESY-HH_LOCALGROUPDISK should go to naf-atlas-support@desy.de, please don't forget to specify your CERN username and how much additional quota is needed.


Similarly, you can check your current list of Rucio rules on each RSE here and request transfers via this interface. The Rucio command line interface is documented here.


The overall use of LOCALGROUPDISKs in the German cloud can be checked using this monitoring (grid certificate required in browser).

Here are some direct links for DESY-HH_LOCALGROUPDISK and DESY-ZN_LOCALGROUPDISK.


Batch system on the NAF


The NAF batch system is part of the BIRD (Batch Infrastructure Resource at DESY) general purpose cluster, based on HTCondor.

Full documentation, including details on OS versions, example configurations can be found on this page.


In particular, there are pages on submitting jobs and checking and managing jobs on the batch system.


Some monitoring is available here.


Problems with the batch system can be reported directly to bird.service@desy.de, including for GPU slots.


Access to NAF GPUs


There are a number of GPUs available within the NAF batch system, information on those is available here.

Access is restricted to people with the nafgpu resource, ask for this via naf-atlas-support@desy.de.


ATLAS software


The NAF has the advantage for ATLAS is that all software is taken from CERN via CVMFS. The following repositories are mounted on all ATLAS NAF working group servers, as well as the batch system:

/cvmfs/atlas.cern.ch
/cvmfs/atlas-condb.cern.ch
/cvmfs/atlas-nightlies.cern.ch
/cvmfs/sft.cern.ch
/cvmfs/sft-nightlies.cern.ch
/cvmfs/unpacked.cern.ch

ATLAS has a very active analysis support, DAST, which is documented here and can be contacted via atlasdast@gmail.com.


Initialising the environment


To set up the ATLAS software, do:

export ATLAS_LOCAL_ROOT_BASE=/cvmfs/atlas.cern.ch/repo/ATLASLocalRootBase
source /cvmfs/atlas.cern.ch/repo/ATLASLocalRootBase/user/atlasLocalSetup.sh

Note that adding these three lines to your .zprofile, in particular the source command, creates some setup interference and causes problems with global PATH variables.

Therefore, it is advisable to execute them only when you have successfully logged in to the NAF.


From this command, you will get an output like this:

lsetup               lsetup <tool1> [ <tool2> ...] (see lsetup -h):
 lsetup agis          ATLAS Grid Information System
 lsetup asetup        (or asetup) to setup an Athena release
 lsetup atlantis      Atlantis: event display
 lsetup eiclient      Event Index 
 lsetup emi           EMI: grid middleware user interface 
 lsetup ganga         Ganga: job definition and management client
 lsetup lcgenv        lcgenv: setup tools from cvmfs SFT repository
 lsetup panda         Panda: Production ANd Distributed Analysis
 lsetup pod           Proof-on-Demand (obsolete)
 lsetup pyami         pyAMI: ATLAS Metadata Interface python client
 lsetup root          ROOT data processing framework
 lsetup rucio         distributed data management system client
 lsetup views         Set up a full LCG release
 lsetup xcache        XRootD local proxy cache
 lsetup xrootd        XRootD data access
advancedTools        advanced tools menu
diagnostics          diagnostic tools menu
helpMe               more help
printMenu            show this menu
showVersions         show versions of installed software

%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
You are running on a RHEL7 compatible OS.
Please refer to these pages for the status and open issues:
 For releases:
  https://twiki.cern.ch/twiki/bin/view/AtlasComputing/CentOS7Readiness#ATLAS_software_status
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%


Diagnostics


To use some diagnostics tools provided by ATLAS, type

 diagnostics

which will return:

 checkOS              check the system OS of your desktop
 db-fnget             run fnget test for Frontier-squid access
 db-readReal          run readReal test for Frontier-squid access
 gridCert             check for user grid certificate issues
 rseCheck             check a Rucio RSE
 runKV                run Kit Validation of an Athena release
 setMeUp              check readiness for tutorial
 setMeUpData          download datafiles for tutorial
 supportInfo          dump information to send to user support
 toolTest             test one or more tools

For example, gridCert is a very useful tool that performs several tests on your grid certificate, in case you fear something is broken.


Some examples


Rucio is initialised as follows:

lsetup rucio

which also issues the lsetup emi command to set up grid tools, allowing you to get a VOMS proxy (see here).


The Athena software environment is initialised in the following way, for example in release 21.2:

asetup AthAnalysis,21.2,latest

If one needs a new version of ROOT, the recommended way is to use LCG views (the full list of LCG releases is available here). One can choose the release, target OS and compiler used, for example:

lsetup "views LCG_100 x86_64-centos7-gcc10-opt"

There is still a possibility to load a specific version of ROOT without other software in the LCG stack, but this is not recommended anymore, and newer releases might not be available:

lsetup "root 6.18.04-x86_64-centos7-gcc8-opt"

ATLAS: WorkBook/NAF2 (last edited 2024-08-07 17:18:28 by DavidSouth)