@@ -13,9 +13,11 @@ try https://variantspark.readthedocs.io/en/latest/getting_started.html
1313 pip3 show variant-spark # To find where variant-spark is installed
1414```
1515- docker cp variantspark_script.py vsrun2:/app/VariantSpark/variantspark_script.py # copy file from local to docker
16- - vs works without mvn install but only pip install inside docker container
17- - variant-spark importance -if gitHub/VariantSpark/data/chr22_1000.vcf -ff gitHub/VariantSpark/data/chr22-labels.csv -fc 22_16050408 -v -rn 500 -rbs 20 -ro -sr 13
16+ - vs works with only ` pip install variant-spark ` :
17+ - ` which variant-spark ` # to find https://github.com/aehrc/VariantSpark/tree/master/bin/variant-spark
18+ - to run scala importance don't need hail to be installed (no mvn install, no pip install hails)
1819 ```
20+ root@16542009db87:/app/VariantSpark# variant-spark importance -if gitHub/VariantSpark/data/chr22_1000.vcf -ff gitHub/VariantSpark/data/chr22-labels.csv -fc 22_16050408 -v -rn 500 -rbs 20 -ro -sr 13
1921 root@16542009db87:/app/VariantSpark# variant-spark importance -if gitHub/VariantSpark/data/chr22_1000.vcf -ff gitHub/VariantSpark/data/chr22-labels.csv -fc 22_16050408 -v -rn 500 -rbs 20 -ro -sr 13
2022 25/10/27 08:41:06 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2123 log4j:WARN No appenders could be found for logger (au.csiro.variantspark.cli.ImportanceCmd).
@@ -29,13 +31,12 @@ try https://variantspark.readthedocs.io/en/latest/getting_started.html
2931 22_16051480_T_C,17.593204808682323
3032 ...
3133 ```
32- - variant-spark --spark --master 'local[*]' -- importance -if gitHub/VariantSpark/data/chr22_1000.vcf -ff gitHub/VariantSpark/data/chr22-labels.csv -fc 22_16050408 -v -rn 500 -rbs 20 -ro -sr 13
33-
34- - ` which variant-spark ` # to find variant-spark bash script
35- - seek bash script: https://github.com/aehrc/VariantSpark/tree/master/bin/variant-spark
36- - it requires to set up spark
37-
34+ - ` python compute_local_fdr.py ` failed with error "ModuleNotFoundError: No module named 'hail'"
35+ - ` python run_importance_chr22.py ` failed with error "ModuleNotFoundError: No module named 'pyspark'"
36+
3837- pip3 install --no-cache-dir -r /app/VariantSpark/requirements.txt # install python dependency
38+ - now hai and pyspark are installed in container: /usr/local/lib/python3.8/dist-packages/hail/
39+ - now ` python compute_local_fdr.py ` and ` python run_importance_chr22.py ` works
3940## pip install variant-spark
4041- it only install compulsary dependency like typecore, but not includes hail and payspark.
4142- /usr/local/lib/python3.8/dist-packages/varspark is installed, includes
0 commit comments