1.安装Scala 并设置环境变量
curl -fL https://github.com/coursier/coursier/releases/latest/download/cs-x86_64-pc-linux.gz | gzip -d > cs && chmod +x cs && ./cs setup --install-dir /share/biosoft/scala/bin
export PATH=/share/biosoft/scala/bin:$PATH
2, 安装Spark 并设置环境变量:
wget https://mirrors.tuna.tsinghua.edu.cn/apache/spark/spark-4.0.1/spark-4.0.1-bin-hadoop3.tgz
tar zxvf spark-4.0.1-bin-hadoop3.tgz
export SPARK_HOME=/share/biosoft/spark/spark-4.0.1-bin-hadoop3
export PATH=$PATH:$SPARK_HOME/bin ./run-example SparkPi # 测试Pi 计算

cp conf/spark-env.sh.template conf/spark-env.sh #添加下面的行
export SCALA_HOME=/share/biosoft/scala/bin/
export JAVA_HOME=/share/biosoft/java/jdk-17.0.9/
export SPARK_HOME=/share/biosoft/spark/spark-4.0.1-bin-hadoop3
gatk --java-options '-Xmx100g' HaplotypeCallerSpark --spark-master local[20] \
-R Arabidopsis_thaliana.TAIR10.dna.chromosome.4.fa \
-I p1.sorted.dedup.bam \
-O p1.g.vcf.gz --max-alternate-alleles 4 --sample-ploidy 2 -ERC GVCF --tmp-dir tmp

如果觉得我的文章对您有用,请随意打赏。你的支持将鼓励我继续创作!
