PacificBiosciences/FALCON

Assembly with example data on slurm failed: {Node(0-rawreads/build)}

Closed this issue · 6 comments

Any feedback please?

I have been struggling for a while to setup Falcon (falcon-kit 1.8.1) on slum to get it run example F1_bull_test data before using Falcon configuration for my Pacbio data.

I have got the F1_bull_test data from here and sadly I have not succeeded it getting it successfully assembly the this F1_bull_test data.

Can you please help me?

Here is the detail below:

pb-assembly suite installed in a separate an environment called 'falcon'

conda create -n falcon
source activate falcon
conda install pb-assembly

Lists of packages within falcon environment

# packages in environment at /home/miniconda3/envs/falcon:
#
# Name                    Version                   Build  Channel
_libgcc_mutex             0.1                 conda_forge    conda-forge
_openmp_mutex             4.5                      1_llvm    conda-forge
bedtools                  2.29.2               hc088bd4_0    bioconda
blas                      1.0                         mkl  
blasr                     5.3.3                h707fff8_0    bioconda
blasr_libcpp              5.3.3                h707fff8_0    bioconda
bwa                       0.7.17               hed695b0_7    bioconda
bzip2                     1.0.8                h516909a_2    conda-forge
ca-certificates           2020.1.1                      0    anaconda
certifi                   2020.4.5.1               py37_0    anaconda
curl                      7.69.1               hbc83047_0  
decorator                 4.4.2                      py_0    conda-forge
falcon-kit                1.8.1                    pypi_0    pypi
falcon-phase              1.2.0                    pypi_0    pypi
falcon-unzip              1.3.7                    pypi_0    pypi
future                    0.18.2           py37hc8dfbb8_1    conda-forge
hdf5                      1.10.4          nompi_h3c11f04_1106    conda-forge
htslib                    1.9                  ha228f0b_7    bioconda
intel-openmp              2020.0                      166    anaconda
krb5                      1.17.1               h2fd8d38_0    conda-forge
ld_impl_linux-64          2.34                 h53a641e_0    conda-forge
libblas                   3.8.0                    15_mkl    conda-forge
libcblas                  3.8.0                    15_mkl    conda-forge
libcurl                   7.69.1               h20c2e04_0  
libdeflate                1.0                  h14c3975_1    bioconda
libedit                   3.1.20181209         hc058e9b_0  
libffi                    3.2.1             he1b5a44_1007    conda-forge
libgcc-ng                 9.2.0                h24d8f2e_2    conda-forge
libgfortran-ng            7.3.0                hdf63c60_5    conda-forge
libgomp                   9.2.0                h24d8f2e_2    conda-forge
liblapack                 3.8.0                    15_mkl    conda-forge
libopenblas               0.3.9                h5ec1e0e_0    conda-forge
libssh2                   1.9.0                h1ba5d50_1  
libstdcxx-ng              9.2.0                hdf63c60_2    conda-forge
llvm-openmp               10.0.0               hc9558a2_0    conda-forge
minimap2                  2.17                 h8b12597_1    bioconda
mkl                       2020.0                      166    conda-forge
mkl-service               2.3.0            py37he904b0f_0  
mkl_fft                   1.0.15           py37ha843d7b_0  
mkl_random                1.1.0            py37hd6b4f25_0  
mummer4                   4.0.0beta2      pl526he1b5a44_5    bioconda
ncurses                   6.1               hf484d3e_1002    conda-forge
networkx                  2.4                        py_1    conda-forge
nim-falcon                1.1.0                h2f8a211_0    bioconda
numpy                     1.18.1           py37h8960a57_1    conda-forge
numpy-base                1.18.1           py37hde5b4d6_1  
openssl                   1.1.1                h7b6447c_0    anaconda
pb-assembly               0.0.8                         0    bioconda
pb-dazzler                0.0.1                h516909a_0    bioconda
pb-falcon                 2.2.4            py37hed50d52_0    bioconda
pb-falcon-phase           0.1.0                h3889886_0    bioconda
pbbam                     0.23.0               h88345a9_0    bioconda
pbgcpp                    1.9.0                         0    bioconda
pbmm2                     1.2.1                h56fc30b_0    bioconda
pcre                      8.44                 he1b5a44_0    conda-forge
perl                      5.26.2            h516909a_1006    conda-forge
pip                       20.0.2                     py_2    conda-forge
pysam                     0.15.3           py37hda2845c_1    bioconda
python                    3.7.6                h0371630_2  
python-edlib              1.3.8.post1      py37hc9558a2_0    bioconda
python-intervaltree       3.0.2                      py_0    bioconda
python-msgpack            0.6.1            py37hc9558a2_1    bioconda
python-sortedcontainers   2.1.0                      py_0    bioconda
python_abi                3.7                     1_cp37m    conda-forge
racon                     1.4.12               he513fc3_0    bioconda
readline                  7.0                  h7b6447c_5  
samtools                  1.9                 h10a08f8_12    bioconda
setuptools                46.1.3           py37hc8dfbb8_0    conda-forge
six                       1.14.0                   py37_0  
sqlite                    3.31.1               h7b6447c_0  
tk                        8.6.10               hed695b0_0    conda-forge
wheel                     0.34.2                     py_1    conda-forge
xz                        5.2.5                h516909a_0    conda-forge
zlib                      1.2.11            h516909a_1006    conda-forge

Contents of fc_run.cfg

[job.defaults]
JOB_QUEUE = bigmem
MB = 30000
NPROC = 6
job_type=slurm

submit = srun --wait=0 -p myqueue  \
    -J ${JOB_NAME}           \
    -o ${JOB_STDOUT}         \
    -e ${JOB_STDERR}         \
    --mem-per-cpu=${MB}M     \
    --cpus-per-task=${NPROC} \
    ${JOB_SCRIPT}
    "${CMD}"

[General]
input_type = raw
#input_type = preads

input_fofn = subreads.fasta.fofn

# The length cutoff used for seed reads used for initial mapping
genome_size=20000000
seed_coverage = 40
length_cutoff = -1

# The length cutoff for preads
length_cutoff_pr = 12000

falcon_greedy = False
falcon_sense_greedy=False

# overlapping options for Daligner
pa_daligner_option = -e0.76 -l1200 -k18 -h480 -w8 -s100
ovlp_daligner_option = -k24 -h480 -e.95 -l1800 -s100
pa_HPCdaligner_option = -v -B128 -M24
ovlp_HPCdaligner_option = -v -B128 -M24

pa_HPCTANmask_option = -k18 -h480 -w8 -e.8 -s100
pa_HPCREPmask_option = -k18 -h480 -w8 -e.8 -s100
#pa_REPmask_code=1,20;10,15;50,10

pa_DBsplit_option = -x500 -s200
ovlp_DBsplit_option = -s400

# error correction consensus option
falcon_sense_option = --output_multi --min_idt 0.70 --min_cov 4 --max_n_read 200 --n_core 24

# overlap filtering options
overlap_filtering_setting = --max_diff 120 --max_cov 120 --min_cov 4 --n_core 24

[job.step.da]
NPROC=4
MB=32000
njobs=300
[job.step.la]
NPROC=8
MB=64000
njobs=200
[job.step.cns]
NPROC=8
MB=64000
njobs=200
[job.step.pda]
NPROC=8
MB=64000
njobs=200
[job.step.pla]
NPROC=4
MB=32000
njobs=300
[job.step.asm]
NPROC=24
MB=192000
njobs=1

contents of fc.246025.err

`"falcon_sense_option=--output_multi --min_idt 0.70 --min_cov 4 --max_n_read 200 --n_core 24". Those should be "-", as in
 "falcon_sense_option=--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24". Auto-replaced.
[WARNING]Option contains flags with "_":
 "overlap_filtering_setting=--max_diff 120 --max_cov 120 --min_cov 4 --n_core 24". Those should be "-", as in
 "overlap_filtering_setting=--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24". Auto-replaced.
[WARNING]Unexpected keys in input config: {'falcon_greedy'}
[WARNING]Please supply a default for "njobs" (aka concurrency) in section [job.defaults]. For now, we will use 8
[INFO]cfg=
{
  "General": {
    "LA4Falcon_preload": false,
    "avoid_text_file_busy": true,
    "bestn": 12,
    "dazcon": false,
    "falcon_greedy": "False",
    "falcon_sense_greedy": false,
    "falcon_sense_option": "--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24",
    "falcon_sense_skip_contained": false,
    "fc_ovlp_to_graph_option": " --min-len 12000",
    "genome_size": "20000000",
    "input_fofn": "subreads.fasta.fofn",
    "input_type": "raw",
    "length_cutoff": "-1",
    "length_cutoff_pr": "12000",
    "overlap_filtering_setting": "--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24",
    "ovlp_DBdust_option": "",
    "ovlp_DBsplit_option": "-s400",
    "ovlp_HPCdaligner_option": "-v -B128 -M24",
    "ovlp_daligner_option": "-k24 -h480 -e.95 -l1800 -s100",
    "pa_DBdust_option": "",
    "pa_DBsplit_option": "-x500 -s200",
    "pa_HPCREPmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCTANmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCdaligner_option": "-v -B128 -M24",
    "pa_REPmask_code": "0,300/0,300/0,300",
    "pa_daligner_option": "-e0.76 -l1200 -k18 -h480 -w8 -s100",
    "pa_dazcon_option": "-j 4 -x -l 500",
    "pa_fasta_filter_option": "streamed-internal-median",
    "pa_subsample_coverage": 0,
    "pa_subsample_random_seed": 12345,
    "pa_subsample_strategy": "random",
    "seed_coverage": "40",
    "skip_checks": false,
    "target": "assembly"
  },
  "job.defaults": {
    "JOB_QUEUE": "bigmem",
    "MB": "30000",
    "NPROC": "6",
    "job_type": "slurm",
    "njobs": 8,
    "pwatcher_type": "fs_based",
    "submit": "srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem
-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n\"${CMD}\"",
    "use_tmpdir": false
  },
  "job.step.asm": {
    "MB": "192000",
    "NPROC": "24",
    "njobs": "1"
  },
  "job.step.cns": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
},
  "job.step.da": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  },
  "job.step.dust": {},
  "job.step.la": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pda": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pla": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  }
}

[INFO]In simple_pwatcher_bridge, pwatcher_impl=<module 'pwatcher.fs_based' from '/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py'>
[INFO]job_type='slurm', (default)job_defaults={'JOB_QUEUE': 'bigmem', 'MB': '30000', 'NPROC': '6', 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}, use_tmpdir=False, squash=False, job_name_style=0
[INFO]Setting max_jobs to 8; was None
[INFO]Num unsatisfied: 2, graph: 2
[INFO]About to submit: Node(0-rawreads/build)
[INFO] starting job Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}) w/ job_type=SLURM
[INFO]$ srun --wait=0 -p myqueue  \
-J Pac163ff68e5c9f           \
-o stdout         \
-e stderr         \
--mem-per-cpu=4000M     \
--cpus-per-task=1 \
/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash
"/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash" >
[ERROR]In pwatcher.fs_based.cmd_run(), failed to submit background-job:
MetaJobSlurm(MetaJob(job=Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}), lang_exe='/bin/bash'))
Traceback (most recent call last):
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 474, in cmd_run
    state.submit_background(bjob)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 121, in submit_background
    bjob.submit(self, exe, script_fn) # Can raise
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 325, in submit
    self.submit_capture = capture(sge_cmd)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/io.py", line 50, in capture
    raise Exception(msg)
Exception: 126 <- 'srun --wait=0 -p myqueue  \\\n-J Pac163ff68e5c9f           \\\n-o stdout         \\\n-e stderr         \\\n--mem-per-cpu=4000M     \\\n--cpus-per-task=1 \\\n/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash\n"/home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash"':
srun: error: cnode84: tasks 0-1: Exited with exit code 13
srun: Terminating job step 246025.0
/bin/sh: line 7: /home/data/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash: Permission denied

[ERROR]Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}
[INFO]CALL:
 scancel -n Pac163ff68e5c9f
Traceback (most recent call last):
  File "/home/miniconda3/envs/falcon/bin/fc_run", line 11, in <module>
    load_entry_point('falcon-kit==1.8.1', 'console_scripts', 'fc_run')()
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 706, in main
    main1(argv[0], args.config, args.logger)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 73, in main1
    input_fofn_fn=input_fofn_fn,
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 235, in run
    dist=Dist(NPROC=4, MB=4000, job_dict=config['job.step.da']),
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/pype.py", line 106, in gen_parallel_tasks
    wf.refreshTargets()
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 278, in refreshTargets
    self._refreshTargets(updateFreq, exitOnFailure)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 323, in _refreshTargets
    raise Exception(msg)
Exception: Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}

`

Thank you,

Thanks Bas Tolhuis from from PacBio techsupport @PacificBiosciences

I got feedback to check 0-rawreads/build for stderr file and the following error message noticed

2020-04-13 11:31:50,050 - root - ERROR - Was waiting for '../../subreads.fasta.fofn'

Checking my files again I found that subreads.fasta.fofn file was missing.

After creating the subreads.fasta.fofn file containing absolute path to the input tutorial F1_bull_test.subreads.fasta.gz file job now submitted to slurm. Progress will be updated.

resubmitted job failed again.

I couldn’t see stderr file in … 0-rawreads/build path.

List of files in in …0-rawreads/build

lrwxrwxrwx 1 sadikm sadikm 108 Apr 20 23:27 pwatcher.dir -> /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/jobs/Pac163ff68e5c9f
-rw-rw-r-- 1 sadikm sadikm 269 Apr 20 23:27 run.sh
-rw-rw-r-- 1 sadikm sadikm 610 Apr 20 23:27 task.sh
-rw-rw-r-- 1 sadikm sadikm 368 Apr 20 23:27 task.json
-rw-rw-r-- 1 sadik sadikm 296 Apr 20 23:27 template.sh

contents of fc.2467347.err

falcon-kit 1.8.1 (pip thinks "falcon-kit 1.8.1")
pypeflow 2.3.0
[INFO]Setup logging from file "None".
[INFO]$ lfs setstripe -c 12 /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data >
[INFO]Apparently '/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data' is not in lustre filesystem, which is fine.
[INFO]fc_run started with configuration fc_run.cfg
[WARNING]Option contains flags with "_":
 "falcon_sense_option=--output_multi --min_idt 0.70 --min_cov 4 --max_n_read 200 --n_core 24". Those should be "-", as in
 "falcon_sense_option=--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24". Auto-replaced.
[WARNING]Option contains flags with "_":
 "overlap_filtering_setting=--max_diff 120 --max_cov 120 --min_cov 4 --n_core 24". Those should be "-", as in
 "overlap_filtering_setting=--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24". Auto-replaced.
[WARNING]Unexpected keys in input config: {'falcon_greedy'}
[WARNING]Please supply a default for "njobs" (aka concurrency) in section [job.defaults]. For now, we will use 8
[INFO]cfg=
{
  "General": {
    "LA4Falcon_preload": false,
    "avoid_text_file_busy": true,
    "bestn": 12,
    "dazcon": false,
    "falcon_greedy": "False",
    "falcon_sense_greedy": false,
    "falcon_sense_option": "--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24",
    "falcon_sense_skip_contained": false,
    "fc_ovlp_to_graph_option": " --min-len 12000",
    "genome_size": "20000000",
    "input_fofn": "subreads.fasta.fofn",
    "input_type": "raw",
    "length_cutoff": "-1",
    "length_cutoff_pr": "12000",
    "overlap_filtering_setting": "--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24",
    "ovlp_DBdust_option": "",
    "ovlp_DBsplit_option": "-s400",
    "ovlp_HPCdaligner_option": "-v -B128 -M24",
    "ovlp_daligner_option": "-k24 -h480 -e.95 -l1800 -s100",
    "pa_DBdust_option": "",
    "pa_DBsplit_option": "-x500 -s200",
    "pa_HPCREPmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCTANmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCdaligner_option": "-v -B128 -M24",
    "pa_REPmask_code": "0,300/0,300/0,300",
    "pa_daligner_option": "-e0.76 -l1200 -k18 -h480 -w8 -s100",
    "pa_dazcon_option": "-j 4 -x -l 500",
    "pa_fasta_filter_option": "streamed-internal-median",
    "pa_subsample_coverage": 0,
    "pa_subsample_random_seed": 12345,
    "pa_subsample_strategy": "random",
    "seed_coverage": "40",
    "skip_checks": false,
    "target": "assembly"
  },
  "job.defaults": {
    "JOB_QUEUE": "bigmem",
    "MB": "30000",
    "NPROC": "6",
    "job_type": "slurm",
    "njobs": 8,
    "pwatcher_type": "fs_based",
    "submit": "srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n\"${CMD}\"",
    "use_tmpdir": false
  },
  "job.step.asm": {
    "MB": "192000",
    "NPROC": "24",
    "njobs": "1"
  },
  "job.step.cns": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.da": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  },
  "job.step.dust": {},
  "job.step.la": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pda": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pla": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  }
}
[INFO]In simple_pwatcher_bridge, pwatcher_impl=<module 'pwatcher.fs_based' from '/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py'>
[INFO]job_type='slurm', (default)job_defaults={'JOB_QUEUE': 'bigmem', 'MB': '30000', 'NPROC': '6', 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}, use_tmpdir=False, squash=False, job_name_style=0
[INFO]Setting max_jobs to 8; was None
[INFO]Num unsatisfied: 2, graph: 2
[INFO]About to submit: Node(0-rawreads/build)
[INFO] starting job Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}) w/ job_type=SLURM
[INFO]$ srun --wait=0 -p myqueue  \
-J Pac163ff68e5c9f           \
-o stdout         \
-e stderr         \
--mem-per-cpu=4000M     \
--cpus-per-task=1 \
/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash
"/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash" >
[ERROR]In pwatcher.fs_based.cmd_run(), failed to submit background-job:
MetaJobSlurm(MetaJob(job=Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue  \\\n-J ${JOB_NAME}           \\\n-o ${JOB_STDOUT}         \\\n-e ${JOB_STDERR}         \\\n--mem-per-cpu=${MB}M     \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}), lang_exe='/bin/bash'))
Traceback (most recent call last):
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 474, in cmd_run
    state.submit_background(bjob)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 121, in submit_background
    bjob.submit(self, exe, script_fn) # Can raise
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 325, in submit
    self.submit_capture = capture(sge_cmd)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/io.py", line 50, in capture
    raise Exception(msg)
Exception: 126 <- 'srun --wait=0 -p myqueue  \\\n-J Pac163ff68e5c9f           \\\n-o stdout         \\\n-e stderr         \\\n--mem-per-cpu=4000M     \\\n--cpus-per-task=1 \\\n/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash\n"/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash"':
srun: error: cnode10: tasks 0-1: Exited with exit code 13
srun: Terminating job step 247347.0
/bin/sh: line 7: /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash: Permission denied

[ERROR]Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}
[INFO]CALL:
 scancel -n Pac163ff68e5c9f
Traceback (most recent call last):
  File "/home/miniconda3/envs/falcon/bin/fc_run", line 11, in <module>
    load_entry_point('falcon-kit==1.8.1', 'console_scripts', 'fc_run')()
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 706, in main
    main1(argv[0], args.config, args.logger)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 73, in main1
    input_fofn_fn=input_fofn_fn,
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 235, in run
    dist=Dist(NPROC=4, MB=4000, job_dict=config['job.step.da']),
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/pype.py", line 106, in gen_parallel_tasks
    wf.refreshTargets()
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 278, in refreshTargets
    self._refreshTargets(updateFreq, exitOnFailure)
  File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 323, in _refreshTargets
    raise Exception(msg)
Exception: Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}

Hi @sadikmu this appears to be a problem I'm running into as well. I believe it is due to the permissions of the .bash scripts that are created in the mypwatcher/wrappers/ directory.

If you change the permissions of the file to be executable and then submit your run script again it will work for that particular step (pertaining to the bash script you just made into an executable). However, it seems we'll have to do this for each individual step & I'd like it to just run. Do any of the PacBio folks have any suggestions/workarounds?

hi @shannonekj,
It would be unrealistic to rerun on slurm queue every time it fails modifying in the file permission to executable which should be autogenerated executable the job is submitted.

The F1_bull_test data (~462Mb) and the fc_run.cfg script here worked successfully on local server. However, running falcon on actual data on the local sever after modifying genome size in fc_run.cfg used above just failed.

I tried to used various parameter options within fc_run.cfg including parameters in #708 -unfortunately it did not work at all and throws the same error shown below. I have been struggling on this for a while and could not figure it out.

Unable to move forward on this and highly appreciate anyones suggestion.

[INFO]Setup logging from file "None".
[INFO]$ lfs setstripe -c 12 /home/assembly/falcon >
[INFO]Apparently '/home/assembly/falcon' is not in lustre filesystem, which is fine.
[INFO]fc_run started with configuration fc_run.cfg
[WARNING]Unexpected keys in input config: set(['ovlp_concurrent_jobs', 'falcon_greedy'])
[INFO]cfg=
{
  "General": {
    "LA4Falcon_preload": false,
    "avoid_text_file_busy": true,
    "bestn": 12,
    "dazcon": false,
    "falcon_greedy": "False",
    "falcon_sense_greedy": false,
    "falcon_sense_option": "--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24",
    "falcon_sense_skip_contained": false,
    "fc_ovlp_to_graph_option": " --min-len 10000",
    "genome_size": "570000000",
    "input_fofn": "subreads.fasta.fofn",
    "input_type": "raw",
    "length_cutoff": "-1",
    "length_cutoff_pr": "10000",
    "overlap_filtering_setting": "--max-diff 100 --max-cov 100 --min-cov 4 --n-core 24",
    "ovlp_DBdust_option": "",
    "ovlp_DBsplit_option": "-s400",
    "ovlp_HPCdaligner_option": "-v -B128 -M24 -k24 -h600 -e.95 -l1800 -s100",
    "ovlp_concurrent_jobs": "288 -s50",
    "ovlp_daligner_option": "-k24 -h480 -e.95 -l1800 -s100",
    "pa_DBdust_option": "",
    "pa_DBsplit_option": "-x500 -s50",
    "pa_HPCREPmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCTANmask_option": "-k18 -h480 -w8 -e.8 -s100",
    "pa_HPCdaligner_option": "-v -B128 -e0.75 -M24 -l1200 -k14 -h256 -w8 -s100 -t16",
    "pa_REPmask_code": "0,300/0,300/0,300",
    "pa_daligner_option": "-e0.75 -l1200 -k14 -h256 -w8 -s100",
    "pa_dazcon_option": "-j 4 -x -l 500",
    "pa_fasta_filter_option": "streamed-internal-median",
    "pa_subsample_coverage": 0,
    "pa_subsample_random_seed": 12345,
    "pa_subsample_strategy": "random",
    "seed_coverage": "40",
    "skip_checks": false,
    "target": "assembly"
  },
  "job.defaults": {
    "JOB_QUEUE": "default",
    "MB": "32768",
    "NPROC": "6",
    "job_type": "local",
    "njobs": "32",
    "pwatcher_type": "blocking",
    "submit": "/bin/bash -c \"${JOB_SCRIPT}\" > \"${JOB_STDOUT}\" 2> \"${JOB_STDERR}\"",
    "use_tmpdir": false
  },
  "job.step.asm": {
    "MB": "192000",
    "NPROC": "24",
    "njobs": "1"
  },
  "job.step.cns": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.da": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  },
  "job.step.dust": {},
  "job.step.la": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pda": {
    "MB": "64000",
    "NPROC": "8",
    "njobs": "200"
  },
  "job.step.pla": {
    "MB": "32000",
    "NPROC": "4",
    "njobs": "300"
  }
}
[INFO]In simple_pwatcher_bridge, pwatcher_impl=<module 'pwatcher.blocking' from '/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pwatcher/blocking.pyc'>
[INFO]job_type='local', (default)job_defaults={'JOB_QUEUE': 'default', 'pwatcher_type': 'blocking', 'use_tmpdir': False, 'MB': '32768', 'job_type': 'local', 'submit': '/bin/bash -c "${JOB_SCRIPT}" > "${JOB_STDOUT}" 2> "${JOB_STDERR}"', 'NPROC': '6', 'njobs': '32'}, use_tmpdir=False, squash=False, job_name_style=0
[INFO]Setting max_jobs to 32; was None
[INFO]Num unsatisfied: 2, graph: 2
[INFO]About to submit: Node(0-rawreads/build)
[INFO]Popen: '/bin/bash -c "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pwatcher/mains/job_start.sh" > "/home/assembly/falcon/0-rawreads/build/run-P593808baa38b0f.bash.stdout" 2> "/home/assembly/falcon/0-rawreads/build/run-P593808baa38b0f.bash.stderr"'
[INFO](slept for another 0.0s -- another 1 loop iterations)
[INFO](slept for another 0.3s -- another 2 loop iterations)
[INFO](slept for another 1.2s -- another 3 loop iterations)
[INFO](slept for another 3.0s -- another 4 loop iterations)
[INFO](slept for another 6.0s -- another 5 loop iterations)
[INFO](slept for another 10.5s -- another 6 loop iterations)
[INFO](slept for another 16.8s -- another 7 loop iterations)
[INFO](slept for another 25.2s -- another 8 loop iterations)
[INFO](slept for another 36.0s -- another 9 loop iterations)
[INFO](slept for another 49.5s -- another 10 loop iterations)
[INFO](slept for another 66.0s -- another 11 loop iterations)
[INFO](slept for another 85.8s -- another 12 loop iterations)
[INFO](slept for another 109.2s -- another 13 loop iterations)
[INFO](slept for another 135.6s -- another 14 loop iterations)
[INFO](slept for another 150.0s -- another 15 loop iterations)
[ERROR]Task Node(0-rawreads/build) failed with exit-code=1
[ERROR]Some tasks are recently_done but not satisfied: set([Node(0-rawreads/build)])
[ERROR]ready: set([])
	submitted: set([])
[ERROR]Noop. We cannot kill blocked threads. Hopefully, everything will die on SIGTERM.
Traceback (most recent call last):
  File "/home/sadik.conda/envs/falcon/bin/fc_run", line 11, in <module>
    load_entry_point('falcon-kit==1.3.0', 'console_scripts', 'fc_run')()
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 706, in main
    main1(argv[0], args.config, args.logger)
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 73, in main1
    input_fofn_fn=input_fofn_fn,
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 235, in run
    dist=Dist(NPROC=4, MB=4000, job_dict=config['job.step.da']),
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/pype.py", line 109, in gen_parallel_tasks
    wf.refreshTargets()
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 278, in refreshTargets
    self._refreshTargets(updateFreq, exitOnFailure)
  File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 362, in _refreshTargets
    raise Exception(msg)
Exception: Some tasks are recently_done but not satisfied: set([Node(0-rawreads/build)])

Apologise @pb-cdunn, Any suggestion please? I got stuck on this.

We've been concentrating on pbipa, which now works really well if you have HiFi data (i.e. ccs reads).

Try using an absolute path for
input_fofn = subreads.fasta.fofn
I think it's having trouble finding that.