Skip to content

Latest commit

 

History

History
75 lines (56 loc) · 1.53 KB

how_to_install.md

File metadata and controls

75 lines (56 loc) · 1.53 KB

How to install bio-pipelines?

Requirements:

Tool Version
Python [3.6, 3.7, 3.8]
Scala [2.11]
Apache-Spark [2.4.3]
sbt [1.4.4]
cd bio_pipelines
python3 -m venv PATH_TO_YOUR_VENV
source PATH_TO_YOUR_VENV/bin/activate

To install with Apache Airflow version 1.10.12:

PYTHON_VERSION is your installed Python version, or run:

PYTHON_VERSION="$(python --version | cut -d " " -f 2 | cut -d "." -f 1-2)"

then:

pip install . --constraint https://raw.githubusercontent.com/apache/airflow/constraints-1.10.12/constraints-${PYTHON_VERSION}.txt

**The DAG should appear either in ${AIRFLOW_HOME}/dags/ or in ~/airflow/dags, but if you have other location of your DAGS, then: **

place cannoli.py in your dags folder:

cp bio_pipelines/cannoli.py /path/to/your/dags/

Go to tools/ - to build BioPackage run:

sbt assembly

The value of the bio_pipe_jar variable should be: the absolute path to the jar stored in:

tools/target/scala-2.11/BioPipeline-assembly-0.2-SNAPSHOT.jar

The value of the bio_pipe_cannoli variable should be: the absolute path to the jar stored in:

tools/lib/cannoli-assembly-spark2_2.11-0.11.0-SNAPSHOT.jar

How to set-up Apache Airflow?

Open a terminal window, run:

airflow scheduler

Open a second terminal window, run:

airflow webserver -p 8080

Go to your browser and go to URL:

localhost:8080