Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
Getting access to LHC Computing Grid. One can find a brief introductory material at the RMKI's getting started page![]()
> grid-proxy-init Here, you will be prompted for your grid password. Or:
> grid-proxy-init -valid 4:00 This is the same, but the authentication will be only valid for 4 hours.
Now you can perform various operations on the grid, e.g. send jobs. You can get info on your authentication (e.g. expiration etc.) by grid-proxy-info . You can destroy your authentication proxy by grid-proxy-destroy .
> myproxy-init Here, you will be prompted for your grid password, and to specify a password attached to your so called job proxy to be created. Or:
> myproxy-init -n Here, you won't be asked to specify a password for your proxy.
Running myproxy-init is necessary when you are running long-term jobs. In this case, the myproxy-init ensures that your jobs still will have authentication even after your interactive proxy (obtained by grid-proxy-init ) has expired. You can get information on your job proxy by myproxy-info . You can destroy your job proxy by myproxy-destroy . Note: if you don't use this, you may not be able to retrieve your job outputs for long-term jobs!
![]()
Requirements = (Member("AFS", other.GlueHostApplicationSoftwareRunTimeEnvironment));
One can apply logical operations to the arguments, like:
Requirements = (Member("AFS", other.GlueHostApplicationSoftwareRunTimeEnvironment) && Member("VO-cms-ORCA_8_13_1", other.GlueHostApplicationSoftwareRunTimeEnvironment));
export LCG_CATALOG_TYPE=lfc(for bash), or setenv LCG_CATALOG_TYPE lfc(for tcsh). The Logical File Name (lfn) is resolved by the Logical File Catalogue (lfc) server. You have to specify this: export LFC_HOST=lfc-cms.cern.ch(for bash), or setenv LFC_HOST lfc-cms.cern.ch(for tcsh). Now you can refer to files with their logical file names, which are site-independent. The logical file names have to be unique. A very convenient way to generate them is to imitate the UNIX type file naming scheme. With the command lfc-mkdir one can make logical directories, like
> lfc-mkdir /grid/cms/user_name/some_directory (All logical file names begin with /grid/virtual_organization . If you are a new user, the directory /grid/virtual_organization/user_name will not exist initially: you have to create it first, if you want an own directory.)
There are various commands to manage logical file names, all beginning with lfc- , like lfc-chmod , lfc-getacl , lfc-setacl , lfc-chown , lfc-ln , lfc-rename , lfc-ls , lfc-rm , lfc-mkdir etc. Refer to the man pages of these commands, but the usage of these should be self-explanatory.
As the given logical directory is ready, one can copy files into it. If you want to store a file in the LCG storage infrastructure permanently, copying a file is not enough: you have to register the file in the logical file catalogue. It is done like:
> lcg-cr --vo cms -d grid100.kfki.hu -l lfn:/grid/cms/alaszlo/destionation/testfile.txt \ file:/afs/kfki.hu/home/alaszlo/source/testfile.txt(Here, the -d destination_Storage_Element is optional, and is used to force the destination Storage Element to be destination_Storage_Element ; otherwise the file may be copied to a Storage Element anywhere in the World.) This command will copy and register your file onto LCG storage system. Once your file is there, you can get a copy of it with lcg-cp e.g. as an input for a job:
> lcg-cp --vo cms lfn:/grid/cms/alaszlo/some_directory/test_file.txt file:$PWD/test_file.txt(This command stages out the file in question onto a local --or AFS-- disc area, pointed at by $PWD , i.e. into your current directory.)
There are various commands to manage the stored files (e.g. unregister and delete them, or create replicas), all beginning with lcg- , like lcg-rep , lcg-aa , lcg-rf , lcg-cp , lcg-cr , lcg-la , lcg-uf , lcg-del , lcg-lg , lcg-lr , lcg-gt , lcg-ra etc. Refer to the man peges of these commands, but most of them are self-explanatory.
Unfortunately, one cannot copy and register complete directory trees. Therefore, I wrote a shell script wrapper to do this:
> lcgcr.sh source_directory destination_directory_lfn (You can find this here: lcgcr.sh.)
Note1: unfortunately, I encountered that the lcg-cr action often fails (this should be remedied in future). Therefore, it is convenient to put an lcg-cr command into a checker loop: | ||||||||
Added: | ||||||||
> > | #!/bin/sh | |||||||
RESULT=1 | ||||||||
Added: | ||||||||
> > | TRIES=0 | |||||||
while (( $RESULT = 0 )) ; do
lcg-cr --vo cms -d grid100.kfki.hu -l lfn:/grid/cms/alaszlo/destintation/test_file.txt file:/afs/kfki.hu/home/alaszlo/source/test_file.txt![]() | ||||||||
Added: | ||||||||
> > | TRIES=$(($TRIES+1)) if (( $TRIES >= 5 )) ; then echo "lcg-cr failed. Giving up after 5 tries." ; exit 1 ; fi | |||||||
if (( $RESULT = 0 )) ; then sleep 1m ; fi
done
This precaution may also be recommended for lcg-cp or other lcg- commands.
Note2: NEVER RESTRICT THE PERMISSIONS OF A FILE OR DIRECTORY IN SUCH A WAY, THAT THE GROUP DOES NOT HAVE WRITE PERMISSIONS. There is a very simple reason for it. It is not really you, who does the actual copying onto a storage element: some program copies it, which is assigned to some more or less random userid, which is mapped to your userid TEMPORARYLY. This is a rather questionable way of manging userid-s, but that is the way how it is implemented in the current version. This fact has rahter dangerous implications. If you restrict the permissions, such that the group does not have write permissions, then at an other occasion (when your userid is mapped to another actual userid), you cannot write your files. :)) Meanwhile, an other user (who is accidentally given that particular userid) may still have write permission to your file, despite of your precaution. :)) Funny, isn't it? As a consequence, noone should restrict the right permissions in such a way that the group does not have write permissions. But this also has a consequence: practically anyone has write permissions to your data, plus you have write permissions for the Nobel prize winner CMS Higgs DST files... Let's hope this conceptional mistake will be corrected soon.
gstream , igstream , ogstream , like the usual C++ STL fstream , ifstream , ofstream file input/output stream classes; the letter 'g' standing for 'grid'). It does nothing else, but treats the file as a normal file, unless its name begins with the string /grid/ . In this case, it stages out the datafile in question onto a local (or AFS) area, and then treats the local file as a normal file. One commonly faces the problem that the file not only has to be processed, but it also has to be passed through a filter programme. Therefore I also wrote pipe streams for grid storage (igpstream , ogpstream ), which are based on the ipstream and opstream classes of the library at http://pstreams.sourceforge.net![]() #include "gstream.h" int main(int argc, char *argv[]) { // Open the datafile for reading. | ||||||||
Changed: | ||||||||
< < | igstream ifile("/grid/cms/alaszlo/some_datafile.dat"); | |||||||
> > | igstream igfile("/grid/cms/alaszlo/some_datafile.dat"); | |||||||
// Extract data from your datafile with 'igstream::operator>>' or with 'igstream::read(char*, int)'. // Close the datafile. | ||||||||
Changed: | ||||||||
< < | ifile.close(); | |||||||
> > | igfile.close(); | |||||||
// Open the datafile for writing. | ||||||||
Changed: | ||||||||
< < | ogstream ofile("/grid/cms/alaszlo/some_datafile.dat"); | |||||||
> > | ogstream ogfile("/grid/cms/alaszlo/some_datafile.dat"); | |||||||
// Write data to your datafile with 'ogstream::operator<<' or with 'ogstream::write(char*, int)'. // Close the datafile. | ||||||||
Changed: | ||||||||
< < | ofile.close(); | |||||||
> > | ogfile.close(); | |||||||
// Open the datafile for reading, through a filter program. | ||||||||
Changed: | ||||||||
< < | igpstream ipfile("/grid/cms/alaszlo/some_datafile.dat.gz", "gunzip --stdout %f"); | |||||||
> > | igpstream igpfile("/grid/cms/alaszlo/some_datafile.dat.gz", "gunzip --stdout %f"); | |||||||
// Extract data from your datafile with 'igpstream::operator>>' or with 'igpstream::read(char*, int)'. // Close the datafile. | ||||||||
Changed: | ||||||||
< < | ipfile.close(); | |||||||
> > | igpfile.close(); | |||||||
// Open the datafile for writing, through a filter program. | ||||||||
Changed: | ||||||||
< < | ogpstream opfile("/grid/cms/alaszlo/some_datafile.dat.gz", "gzip - > %f"); | |||||||
> > | ogpstream ogpfile("/grid/cms/alaszlo/some_datafile.dat.gz", "gzip - > %f"); | |||||||
// Write data to your datafile with 'ogpstream::operator<<' or with 'ogpstream::write(char*, int)'. // Close the datafile. | ||||||||
Changed: | ||||||||
< < | opfile.close(); | |||||||
> > | ogpfile.close(); | |||||||
return 0;
}
You can get the library files here: pstream.h, gstream.h, gstream.cc. Before you can use it you have to export the following environmental variables:
export VO=your_vo(for bash), or setenv VO your_vo(for tcsh), and export DEST=your_favourite_storage_element(for bash), or setenv DEST your_favourite_storage_element(for tcsh). Setting the environmental variable TMPDIR is optional. This specifies the local (or AFS) directory, where the datafiles are staged out (therefore, it has to have large disk space!). E.g.:
export TMPDIR=/tmp(for bash), or setenv TMPDIR /tmp(for tcsh). If not specified, the current working directory ( $PWD ) is used, as this is recommended for grid jobs (the working nodes have large disk spaces).
-- AndrasLaszlo - 22 Mar 2006 | ||||||||
Changed: | ||||||||
< < |
| |||||||
> > |
| |||||||
| ||||||||
Changed: | ||||||||
< < |
| |||||||
> > |
| |||||||
|