CEP quick-start (including example run)

This section provides some quick notes on getting started with the pipeline system. More details are available in subsequent sections of this chapter. The first chapter detail all the steps needed to start an imaging pipeline from scratch until checking the output meta-data. (test data INCLUDED!!)

Setting up the environment and directories

The pipelines and framework you will be using are exactly the same as the automated central processing system. To allow usage on your own data some configuration has to be done. Keep in mind most of these steps have to be performed only a single time

Step by Step:

  1. Log in on lfe001, the head-node of the cep1 cluster. The pipelines should only be started on this cluster: The resource usage can be large and might interfere with observation!.

  2. Load an environment:

    1. use LofIm for the latest development version use Pythonlibs for the external python packages
  3. Create directories:

    1. cexec lce: "mkdir /data/scratch/USERNAME Create a personal directory on the computation nodes.

      Fill in your own user-name.

    2. mkdir -p /home/USERNAME/pipeline/runtime_directory Create in your home a directory for runtime files.

    3. mkdir /home/USERNAME/pipeline/config_files Create a directory for the config files.

    4. mkdir /home/USERNAME/pipeline/parset_files Create a directory for the parset files.

  4. Copy the configuration files to your own config dir:

    1. The enviroment used has influence on the location of some files. LofIm for instance is renewed each day with the resulting file in a different directory. Locate the current directory for your invironment. which msss_calibrator_pipeline.py This could result in the folowing: /opt/cep/LofIm/daily/Fri/lofar_build/install/gnu_opt/bin/msss_calibrator_pipeline.py The config files are found relative to the path install: install/gnu_opt/share/pipeline
    2. Copy the pipeline.cfg and tasks.cfg files to your own configuration directory. Take the first part of the found path and add gnu_opt/share/pipeline copy all cfg files to your own config location. cp /opt/cep/lofar/lofar_versions/LOFAR-Release-1_3-latest/lofar_build/install/gnu_opt/share/pipeline/*.cfg /home/USERNAME/pipeline/config_files
    3. cp /home/klijn/cep1.clusterdesc /home/USERNAME/pipeline/config_files/cep1.clusterdesc Copy the cluster description file to your config dir. It is currentlylocated in a home directory
    4. This copy action will change the dynamic nature of the files. If you want to be sure that you have the bleeding edge software perform this copy step and the next adaptation step again.
  5. Adapt the configuration files so they point to your own directories:

    1. Open your own version pipeline.cfg with your editor of choice.

    2. Observe that the first entry lofarroot points to a daily build -or- a release version. This is the reasoning behind the dynamic nature of the configuration files. And a possible copy if you want to use the latest version.

    3. Change the runtime_directory entry to /home/USERNAME/pipeline/runtime_directory/

      THIS RUNTIME_DIRECTORY MUST BE ACCESSIBLE FROM ALL NODES

    4. Change the working_directory entry to /data/scratch/USERNAME

      THIS WORKING_DIRECTORY CAN –NOT– EVER, FOR ANY REASON, BE ON A GLOBAL SHARE. EVER

    5. Change the clusterdesc entry to /home/USERNAME/pipeline/config_files/cep1.clusterdesc

    6. Change the task_files entry to [/home/USERNAME/pipeline/config_files/tasks.cfg]

#Example pipeline.cfg
[DEFAULT]
lofarroot = /opt/cep/LofIm/daily/Fri/lofar_build/install/gnu_opt
casaroot = /opt/cep/LofIm/daily/Fri/casacore
pyraproot = /opt/cep/LofIm/daily/Fri/pyrap
hdf5root = /opt/cep/hdf5
wcsroot = /opt/cep/wcslib
pythonpath = /opt/cep/LofIm/daily/Fri/lofar_build/install/gnu_opt/lib/python2.6/dist-packages
runtime_directory = /home/klijn/pipeline/runtime_directory
recipe_directories = [%(pythonpath)s/lofarpipe/recipes]
working_directory = /data/scratch/klijn
task_files = [/home/klijn/pipeline/config/tasks.cfg]

[layout]
job_directory = %(runtime_directory)s/jobs/%(job_name)s

[cluster]
clusterdesc = /home/klijn/pipeline/config/cep1.clusterdesc

[deploy]
engine_ppath = %(pythonpath)s:%(pyraproot)s/lib:/opt/cep/pythonlibs/lib/python/site-packages
engine_lpath = %(lofarroot)s/lib:%(casaroot)s/lib:%(pyraproot)s/lib:%(hdf5root)s/lib:%(wcsroot)s/lib

[logging]
log_file = %(runtime_directory)s/jobs/%(job_name)s/logs/%(start_time)s/pipeline.log
  1. Run a short template run of the imaging pipeline:

    1. use LofIm

    2. cp /data/scratch/klijn/out.parset /home/USERNAME/pipeline/parset_files/out.parset copy the test parametersets file to your own parset directory.

    3. `` msss_imager_pipeline.py /data/scratch/USERNAME/out.parset –config ~/pipeline/config_files/pipeline.cfg –job test1 -d`` details:

      1. msss_imager_pipeline.py the imaging pipeline executable
      2. /home/USERNAME/pipeline/parset_files/out.parset the settings for the pipeline
      3. --config ~/pipeline/config_files/pipeline.cfg the configuration to use
      4. --job test1 a self chosen name allows distinguishing between runs
      5. -d turn on debugging information prints. The default settings of the pipeline is almost silent. This settings allows some sense of progress.
      6. The pipeline should now perform a simple imaging run of a msss like observation.
    4. The resulting image can be found at lce001:/data/scratch/USERNAME/test1/awimage_cycle_0

  2. Additional information:

    1. The pipeline remembers progress: And will not redo work already done.
    2. cd /home/USERNAME/pipeline/runtime_directory/jobs/test1 Go to the runtime_directory for the started/finished run. At this location you can find the logs, partial parset, mapfiles (internal datamember) and the statefile.
    3. deleting this state file will reset the pipeline and allows running from the start. (You could also rename your job)
    4. In the parset directory additional parsets will become available. Currently a full mom_parset.parset is provided. It contains ALL settings that are set from outside the pipeline framework.
  3. TODO:

    1. A description of parameter set entries
    2. How-to use your own data
    3. How-to change the executables (task.cfg file changes)
    4. How-to use your own build of the offline processing framework

Table Of Contents

Previous topic

Installation and setup

Next topic

Framework installation

This Page