This shows how to submit a job an EMR (Elastic Map Reduce) cluster on AWS to get id
>>> ./EHPC-EMR --submit -d=DOMAIN --command=COMMAND OPTIONS
All parameters should be in format parameter=value
--command original command to run
--input-files, -i list of input files seprated by commas
--output-files, -o list of expected output files seprated by commas
--domain, -d Domain of the main node
--cache-files, -cf list of cache files to cache to mappers and reducers, for hadoop mode only
--cache-archives, -ca list of cache archives to cache to mappers and reducers, for hadoop mode only
--files list of files to pack with the job
--reducer path of the reducer to execute e.g 'cat', default NONE
--output-dir path of the output dir for the mappers and reducers, default /home/hadoop/output/ID
--conf set of hadoop configuration parameters seprated by commas, for hadoop mode only
--owner the owner of job
if owner is system, the commad will execute on the command line, client will wait the job is Done
if owner is hadoop, the job will be submitted as a Hadoop job
if owner is otherwise, this will be a PBS Torque Job
--no-fetch-output-files Don't fetch output files, in case of s3://
Example 1
Run crossbow on ec2......com
./EHPC-EMR --submit -d=ec2......com -id=1 --command='export CROSSBOW_HOME=/home/hadoop/crossbow;$CROSSBOW_HOME/cb_hadoop --preprocess --input=s3://eg.nubios.us/crossbow/example/hg18/reads.manifest --output=s3://eg.nubios.us/crossbow/example/hg18/output --reference=s3://eg.nubios.us/crossbow-refs/hg18.jar --all-haploids --tempdir=/mnt/tmp --streaming-jar=/home/hadoop/contrib/streaming/hadoop-streaming-0.20.205.jar --just-align' -owner=user1
Returns JOBID