Difference: ArCondNevis (3 vs. 4)

Revision 42010-09-16 - EvanWulf

Line: 1 to 1
 
META TOPICPARENT name="UsingNevisT3"

Using ArCond at Nevis

Line: 8 to 8
  In my experience, a job that normally takes > 1 hour or so to run locally is worth submitting to ArCond (will take ~10 minutes to finish on ArCond). Also, if you are submitting many jobs over many different datasets, writing scripts to submit these all to ArCond rather than running them all sequentially will probably be much faster since you have 48 cores on the worker nodes vs. 16 interactively (that are more consistently in use by others).
Changed:
<
<

Teh Tutorial

>
>

The Tutorial

  This tutorial works in zsh. I'm not sure about bash. It will teach you how to submit c++ jobs which run over D3PDs (in the AnalysisUtilities framework) using ArCond. I'm fairly new to ArCond, so it is possible that I've made mistakes (or that things could be done in a more efficient way). If you discover any, please let me know or update this page yourself.
Line: 31 to 31
 Also, you need to make sure your files are on xrootd. To copy a dataset to xrootd, after doing the above setup, do:
Changed:
<
<
source /data/users/common/xrootadd.sh
>
>
source /data/users/common/xrootadd.sh <destination/directory>
 

If you just now copied a dataset, you'll need to wait a few hours for the database to update before continuing (otherwise, ArCond won't know that the data is available on the nodes by default). To check if your data is there, again do the above setup and do:

Line: 71 to 71
  In the ArCondNevis/arc_d3pd directory, there is a file called arcond.conf. The only three important lines begin with input_data - where you specify the dataset (always in the form /data/xrootd/), max_jobs_per_node (remember there are 3 nodes, so multiply this number by 3 and you'll get the degree of parallelization of your jobs), and package_dir - where you specify the path to your analysis package to be copied to where your jobs will run. Modify these as you see fit (if you just want to run ArCond out-of-the-box for the tutorial, leave these as they are).
Changed:
<
<
Now check out the ArCondNevis/patterns directory. Here you tell ArCond what machines are available (and any requirements for those machines). The files all have the form schema.site.xeniaXX.nevis.columbia.edu. You don't need to modify these, but if nodes are ever added to the T3 site, you'll need to add a corresponding file. One thing you might want to modify is uncommenting the email notification line (else you'll get a ton of emails when the jobs finish... personally i let the emails come and then filter the emails, but this is really up to you).
>
>
Now check out the ArCondNevis/arc_d3pd/patterns directory. Here you tell ArCond what machines are available (and any requirements for those machines). The files all have the form schema.site.xeniaXX.nevis.columbia.edu. You don't need to modify these, but if nodes are ever added to the T3 site, you'll need to add a corresponding file. One thing you might want to modify is uncommenting the email notification line (else you'll get a ton of emails when the jobs finish... personally i let the emails come and then filter the emails, but this is really up to you).
  Finally, check out the ArCondNevis/user directory. There should be 3 files. The most important is called ShellScript_BASIC.sh. Open this and skip to the part where it says "user-defined part" (everything before this is ArCond set up, e.g. copying packages to the nodes, setting up the parallelization, etc.). As you can see, it does some setup for AnalysisUtilities, then compiles the packages, then runs the job with the following line:
 
This site is powered by the TWiki collaboration platform Powered by PerlCopyright © 2008-2020 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki? Send feedback