Welcome to GRID_LRT’s documentation!

This package is built by Alexandar Mechev and the LOFAR e-infra group at Leiden University with the support of SURFsara. The goals of this package is to enable High Throughput processing of LOFAR data on the Dutch GRID infrastructure. We do this by making a set of tools designed to wrap around several different LOFAR processing strategies. These tools are responsible for staging data at the LOFAR Long Term Archives, creating and launching GRID jobs, as well as managing intermediate data on the GRID storage.

Installation

Via Python Package Index

Install the package (or add it to your requirements.txt file):

pip install GRID_LRT

Via Git or Download

Download the latest version from https://www.github.com/apmechev/GRID_LRT. To install, use

python setup.py build
python setup.py install

In the case that you do not have access to the python system libraries, you can use --prefix= to specify install folder. For example if you want to install it into a folder you own (say /home/apmechev/software/python) use the following command:

python setup.py build
python setup.py install --prefix=${HOME}/software/python

Note

NOTE: you need to have your pythonpath containing

“${HOME}/software/python/lib/python[2.6|2.7|3.4]/site_packages”

and that folder needs to exist beforehand or setuptools will complain

Tokens

The GRID_LRT.Token module is responsible for interactions with CouchDB using the PiCaS token framework. It contains a Token_Handler object which manages a single _design document on CouchDB, intended for a set of jobs that are logically linked together. In the LOFAR Surveys case, this holds the jobs of a single Observation. Additionally a Token_Set object can create batch tokens, upload attachments to them in bulk and change Token fields in bulk as well. This module is used in combination with the srmlist class to automatically create sets of jobs with N files each.

Token.py

Location: GRID_LRT/Token.py Imports:

>>> from GRID_LRT.Token import Token_Handler
>>> from GRID_LRT.Token import TokenSet

TokenHandler

TokenSet

Staging Modules

These modules are located in GRID_LRT.Staging and can be used to batch stage or check the status of the files on the GRID Storage.

GRID_LRT.Staging.srmlist

GRID_LRT.Staging.srmlist.count_files_uberftp(directory)[source]
GRID_LRT.Staging.srmlist.make_srmlist_from_gsiftpdir(gsiftpdir)[source]
GRID_LRT.Staging.srmlist.slice_dicts(srmdict, slice_size=10)[source]

Returns a dict of lists that hold 10 SBNs (by default). Missing Subbands are treated as empty spaces, if you miss SB009, the list will include 9 items from SB000 to SB008, and next will start at SB010

class GRID_LRT.Staging.srmlist.srmlist(checkOBSID=True, link=None)[source]

Bases: list

The srmlist class is an extension of Python lists that can hold a list of srm links to data on GRID Storage (LOFAR Archive, Intermediate Storage, etc).

In addition to the regular list capabilities, it also has internal checks for the location and the OBSID of the data. When a new item is appended, these checks are done automatically. Checking OBSID is an optional argument set to True by default.

__init__(checkOBSID=True, link=None)[source]

x.__init__(…) initializes x; see help(type(x)) for signature

append(item)[source]

L.append(object) – append object to end

check_location(item)[source]
check_obsid(item)[source]
check_str_location(item)[source]
count(value) → integer -- return number of occurrences of value
extend()

L.extend(iterable) – extend list by appending elements from the iterable

Returns a generator that can be used to generate links that can be staged/stated with gfal

gfal_replace(item)[source]

For each item, it creates a valid link for the gfal staging scripts

Returns a generator which can be iterated over, this generator will return a set of gsiftp:// links which can be used with globus-url-copy and uberftp

gsi_replace(item)[source]

Returns a generator that can be used to generate http:// links that can be downloaded using wget

http_replace(item)[source]
index(value[, start[, stop]]) → integer -- return first index of value.

Raises ValueError if the value is not present.

insert()

L.insert(index, object) – insert object before index

pop([index]) → item -- remove and return item at index (default last).

Raises IndexError if list is empty or index is out of range.

remove()

L.remove(value) – remove first occurrence of value. Raises ValueError if the value is not present.

reverse()

L.reverse() – reverse IN PLACE

sbn_dict(pref='SB', suff='_')[source]

Returns a generator that creates a pair of SBN and link. Can be used to create dictionaries

sort()

L.sort(cmp=None, key=None, reverse=False) – stable sort IN PLACE; cmp(x, y) -> -1, 0, 1

srm_replace(item)[source]
stringify_item(item)[source]
trim_spaces(item)[source]

Sometimes there are two fields in the incoming list. Only take the first as long as it’s fromatted properly

GRID_LRT.Staging.stage_all_LTA

GRID_LRT.Staging.stage_all_LTA.get_stage_status(stageid)[source]
GRID_LRT.Staging.stage_all_LTA.location(filename)[source]
GRID_LRT.Staging.stage_all_LTA.main(filename, test=False)[source]
GRID_LRT.Staging.stage_all_LTA.process(urls, repl_string, match, test=False)[source]
GRID_LRT.Staging.stage_all_LTA.process_surl_line(line)[source]

Used to drop empty lines and to take the first argument of the srmfile (the srm:// link)

GRID_LRT.Staging.stage_all_LTA.replace(file_loc)[source]
GRID_LRT.Staging.stage_all_LTA.return_srmlist(filename)[source]
GRID_LRT.Staging.stage_all_LTA.state_dict(srm_dict)[source]
GRID_LRT.Staging.stage_all_LTA.strip(item)[source]

GRID_LRT.Staging.state_all

Python module to check the state of files using gfal and return their locality # ===================================================================== # # author: Ron Trompert <ron.trompert@surfsara.nl> – SURFsara # # helpdesk: Grid Services <grid.support@surfsara.nl> – SURFsara # # # # usage: python state.py # # description: # # Display the status of each file listed in “files”. The paths # # should have the ‘/pnfs/…’ format. Script output: # # ONLINE: means that the file is only on disk # # NEARLINE: means that the file in only on tape # # ONLINE_AND_NEARLINE: means that the file is on disk # # and tape # # ===================================================================== #

GRID_LRT.Staging.state_all.check_status(surl_link, verbose=True)[source]

Obtain the status of a file from the given surl.

Args:
param surl:the SURL pointing to the file.
type surl:str
parame verbose:print the status to the terminal.
type verbose:bool
Returns:
(filename, status):
 a tuple containing the file and status as stored in the ‘user.status’ attribute.
GRID_LRT.Staging.state_all.check_status_file(surl_list)[source]

Unimplemented task

GRID_LRT.Staging.state_all.load_file_into_srmlist(filename)[source]

Helper function that loads a file into an srmlist object (will be added to the actual srmlist class later)

GRID_LRT.Staging.state_all.main(filename, verbose=True)[source]

Main function that takes in a file name and returns a list of tuples of filenames and staging statuses. The input file can be both srm:// and gsiftp:// links.

Args:
param filename:The filename holding the links whose have to be checked
type filename:str
param verbose:A toggle to turn off printing out the status of each file.

True by default will print everything out :type verbose: bool

Returns:
ret results:A list of tuples containing the file_name and the State

Usage:

>>> from GRID_LRT.Staging import state_all
>>> filename='/home/apmechev/GRIDTOOLS/GRID_LRT/GRID_LRT/tests/srm_50_sara.txt'
>>> results=state_all.main(filename)
>>> results=state_all.main(filename, verbose=False)
>>> results[0]
('L229507_SB150_uv.dppp.MS_f6fc7fc5.tar', 'ONLINE_AND_NEARLINE')
GRID_LRT.Staging.state_all.percent_staged(results)[source]

Takes list of tuples of (srm, status) and counts the percentage of files that are staged (0->1) and retunrs this percentage as float

Usage:

>>> from GRID_LRT.Staging import state_all
>>> filename='/home/apmechev/GRIDTOOLS/GRID_LRT/GRID_LRT/tests/srm_50_sara.txt'
>>> results=state_all.main(filename, verbose=False)
>>> state_all.percent_staged(results)

GRID_LRT.Staging.stager_access

It uses an xmlrpc proxy to talk and authenticate to the remote service. Your account credentials will be read from the awlofar catalog Environment.cfg, if present or can be provided in a .stagingrc file in your home directory.

!!Please do not talk directly to the xmlrpc interface, but use this module to access the provided functionality. !! This is to ensure that when we change the remote interface, your scripts don’t break and you will only have to upgrade this module.

GRID_LRT.Staging.stager_access.stage(surls)[source]

Stage list of SURLs or a string holding a single SURL

Parameters:surls (either a list() or a str()) – Either a list of strings or a string holding a single surl to stage
Returns:An integer which is used to refer to the stagig request when polling

the API for a staging status

GRID_LRT.Staging.stager_access.get_status(stageid)[source]

Get status of request with given ID

Args:
param stageid:The id of the staging request which you want the status of
type stageid:int
Returns:
status:A string describing the staging status: ‘new’, ‘scheduled’,

‘in progress’ or ‘success’

GRID_LRT.Staging.stager_access.get_surls_online(stageid)[source]

Get a list of all files that are already online for a running request with given ID

GRID_LRT.Staging.stager_access.get_srm_token(stageid)[source]

Get the SRM request token for direct interaction with the SRM site via Grid/SRM tools

GRID_LRT.Staging.stager_access.reschedule(stageid)[source]

Reschedule a request with a given ID, e.g. after it was put on hold due to maintenance

GRID_LRT.Staging.stager_access.get_progress()[source]

Get a detailed list of all running requests and their current progress. As a normal user, this only returns your own requests.

GRID_LRT.Staging.stager_access.get_storage_info()[source]

Get storage information of the different LTA sites, e.g. to check available disk pool space. Requires support role permissions.

Sandbox Module

The Sandbox module creates a tar archive of the scripts to be distributed to the worker nodes at the launch of a PiCaS job. The location of the sandbox is stored in the PiCaS token and upon launch, it is downloaded and extracted. The sandbox is created from a configuration file which defines its name, location scrts repository and any additional processing scripts, such as prefactor.

GRID_LRT.sandbox

Sandbox building and uploading module

class GRID_LRT.sandbox.Sandbox(cfgfile=None, **kwargs)[source]

Bases: object

A set of functions to create a sandbox from a configuration file. Uploads to grid storage and ssh-copies to a remote ftp server as a fallback location.

Usage with a .cfg file:

>>> from GRID_LRT import sandbox
>>> s=sandbox.Sandbox()
>>> s.build_sandbox('GRID_LRT/data/config/bash_file.cfg')
>>> s.upload_sandbox()

This will build the sandbox according to the recipe in bash_file.cfg and upload it to grid storage

__init__(cfgfile=None, **kwargs)[source]

Creates a ‘sandbox’ object which builds and uploads the sanbox. An optional argument is the configuration file which is a yaml file specifying the repositories to include, the type of the sanbox, and its name.

Example configuration files are included in GRID_LRT/data/config.

Parameters:cfgfile (str) – The name of the configuration file to build a sandbox from
build_sandbox(sbx_config)[source]

A comprehensive function that builds a Sandbox from a configuration file and creates a sandbox tarfile.

check_token()[source]

This function does the necessary linkage between the sandbox and token most importantly it saves the tokvar.cfg file in the sbx, but also checks if the token variables are all existing. If so, tokvar is created and put inside the SBX

cleanup()[source]
copy_base_scripts(basetype=None)[source]

Backwards compatible

copy_git_scripts()[source]

Reads the location of the sandbox base scripts repository and clones in the current directory. Checks out the appropriate branch

create_sbx_folder()[source]

Makes an empty sandbox folder or removes previous one

delete_gsi_sandbox(sbxfile)[source]
delete_sbx_folder()[source]

Removes the sandbox folder and subfolders

enter_sbx_folder(directory=None)[source]

Changes directory to the (temporary) sandbox folder)

get_result_loc()[source]
load_git_scripts()[source]

Loads the git scripts into the sandbox folder. Top dir names are defined in the yaml, not by the git name

make_tokvar_dict()[source]
parseconfig(yamlfile)[source]

Helper function to parse the sandbox configuration options from the yaml .cfg file. Loads the options in a dictionary stored in an internal variable

Parameters:yamlfile (str) – The name of the sandbox configuration file
sandbox_exists(sbxfile)[source]
upload_gsi_sbx(loc=None, upload_name=None)[source]

Uploads the sandbox to the relative folders

upload_sandbox(upload_name=None)[source]
upload_sbx(loc=None, upload_name=None)[source]

Uploads sandbox to all possible locations

upload_ssh_sandbox(upload_name=None)[source]
zip_sbx(zipname=None)[source]
class GRID_LRT.sandbox.UnauthorizedSandbox(*args, **kw)[source]

Bases: GRID_LRT.sandbox.Sandbox

__init__(*args, **kw)[source]

Creates a ‘sandbox’ object which builds and uploads the sanbox. An optional argument is the configuration file which is a yaml file specifying the repositories to include, the type of the sanbox, and its name.

Example configuration files are included in GRID_LRT/data/config.

Parameters:cfgfile (str) – The name of the configuration file to build a sandbox from
build_sandbox(sbx_config)

A comprehensive function that builds a Sandbox from a configuration file and creates a sandbox tarfile.

check_token()

This function does the necessary linkage between the sandbox and token most importantly it saves the tokvar.cfg file in the sbx, but also checks if the token variables are all existing. If so, tokvar is created and put inside the SBX

cleanup()
copy_base_scripts(basetype=None)

Backwards compatible

copy_git_scripts()

Reads the location of the sandbox base scripts repository and clones in the current directory. Checks out the appropriate branch

create_sbx_folder()

Makes an empty sandbox folder or removes previous one

delete_gsi_sandbox(sbxfile)
delete_sbx_folder()

Removes the sandbox folder and subfolders

enter_sbx_folder(directory=None)

Changes directory to the (temporary) sandbox folder)

get_result_loc()
load_git_scripts()

Loads the git scripts into the sandbox folder. Top dir names are defined in the yaml, not by the git name

make_tokvar_dict()
parseconfig(yamlfile)

Helper function to parse the sandbox configuration options from the yaml .cfg file. Loads the options in a dictionary stored in an internal variable

Parameters:yamlfile (str) – The name of the sandbox configuration file
sandbox_exists(sbxfile)
upload_gsi_sbx(loc=None, upload_name=None)

Uploads the sandbox to the relative folders

upload_sandbox(upload_name=None)
upload_sbx(loc=None, upload_name=None)

Uploads sandbox to all possible locations

upload_ssh_sandbox(upload_name=None)
zip_sbx(zipname=None)

Error Codes

Here are a list of errors that the GRID_Sandbox or GRID_Launcher return when processing data on a worker node. The error code is saved in the ‘output’ field of the PiCaS token.

-2 -> Sandbox downloaded but size 0kB
-1 ->
0 -> RUN OK!
1 -> One of Token=${TOKEN}, Picas_usr=${PICAS_USR}, Picas_db=${PICAS_DB} not set
2 ->
3 -> Parset doesn’t exist
4 ->
5 ->
6 ->
7 ->
8 ->
9 ->
10 -> Softdrive not found
11 -> LOFAR env cannot be found by GRID_PiCaS_Launcher
12 -> No init_env script
13 ->
14 ->
15 ->
16 ->
17 ->
18 ->
19 ->
20 -> No download File Present
21 -> Download fails
22 -> Data not staged
23 -> pref_cal1 solutions do not download/extract
24 ->
25 ->
26 ->
27 ->
28 ->
29 ->
30 -> No files in uploads folder
31 -> Upload to gsiftp fails
32 -> Upload to gsiftp fails: Pools full!
33 -> Upload to gsiftp fails: File already exists
34 -> Upload to gsiftp fails: File cannot be found (Parent folder not exist?)
35 ->
36 ->
37 ->
90 -> genericpipeline.py stdout file cannot be found!
91 ->
92 ->
93 ->
94 ->
95 ->
96 -> Files not downloaded fully
97 -> dppp memory error in prefactor
98 -> Bad_alloc error in prefactor
99 -> Generic Prefactor Failure


Indices and tables