Holiday Notice:
The Frontline Support team will be slow to respond December 17-18 due to an institute-wide retreat and offline December 22- January 1, while the institute is closed. Thank you for your patience during these next few weeks. Happy Holidays!

Error when running CNVDiscoveryPipeline using the latest version(SVToolkit version 2.00 (build 1747)

zzqzzq ChinaMember
edited August 2017 in GenomeSTRiP

Dear @bhandsaker

I have finished running the deletion discovery pipeline, but when running CNVDiscoveryPipeline using following command, it gave me an error. It is difficult for me to figure it out. Hope your help. Thanks.


java -cp ${classpath} ${mx} \
    org.broadinstitute.gatk.queue.QCommandLine \
    -S ${SV_DIR}/qscript/discovery/cnv/CNVDiscoveryPipeline.q \
    -S ${SV_DIR}/qscript/SVQScript.q \
    -gatk ${SV_DIR}/lib/gatk/GenomeAnalysisTK.jar \
    -cp ${classpath} \
    -tempDir ${SV_TMPDIR}/tmp \
    -configFile /home/conf/genstrip_parameters.txt \
    -R  /home/conf/merge.fa \
    -genderMapFile  /home/conf/gender \
    -runDirectory ${runDir} \
    -md  /home/CNV/$1/metadata/ \
    -jobLogDir ${runDir}/logs \
    -I  /home/conf/bam.list \
    -intervalList  /home/conf/$1.list \
    -ploidyMapFile  /home/conf/ploidymap.txt \
    -jobRunner Shell -gatkJobRunner Shell -tilingWindowSize 5000 \
    -tilingWindowOverlap 2500 -maximumReferenceGapLength 2500 -boundaryPrecision 200 -minimumRefinedLength 2500 -run \
    || exit 1

Error information

Error: Exception processing cnp: null
CNP: CNV_1_35000000_35005000 1:35000001-35005000
##### ERROR ------------------------------------------------------------------------------------------
##### ERROR stack trace 

Best wishes


  • bhandsakerbhandsaker Member, Broadie, Moderator admin

    I am pretty sure this error is due to a bug that occurs when you are not using any mask files. A workaround should be to create a dummy mask file that does not mask any portions of the reference genome and pass that explicitly as a mask file, e.g. using -genomeMaskFile.

    This problem has since been fixed in the code base, so as soon as we make a new release the workaround should not be necessary.

Sign In or Register to comment.