Python versions are a nightmare. You probably know this. Crispor pretty much requires Python 3.9 now since V5.2, because of the rs3 scores. That being said, everything else works with Python 3.6. Up to until 5.2, Python 2.7 was fine. You will probably want a virtual environment. If you are on RedHat/CentOS/Rocky, you can get Python 3.9 from your favorite rpm server, and for Ubuntu it's the same. Compiling Python yourself is also not a big problem.
The file requirements.txt has all version numbers of all packages on crispor.org, but it should not be needed.
In this section, I assume that you are root and you want to setup a local CRISPOR website. If you only want to use the command line tools, the installation commands below would be the same, but 1) you don't need the sudo commands for pip and 2) you can use the option '--user' when running pip to install the tools into your own home directory ~/.local instead of /usr/ and /var/.
If you are unsure what the things below mean or if you just want to try it and not install it or modify your server setup, you may want to try the virtual machine, which is a complete installation of CRISPOR with everything included: http://crispor.org/downloads/
CRISPOR uses python3.9 since version 5.2. Change pip
to pip3
in the commands below if your default python is not python3.
Let me know if you cannot upgrade to Python3.
First, install BWA and a few required basic python modules using your linux package manager:
# Debian/Ubuntu
add-apt-repository ppa:deadsnakes/ppa
apt install python3.9
apt-get install bwa python-pip python-matplotlib
or
# Fedora/Centos/Redhat/Scientific Linux
# find a python3.9 rpm package -> email me if in trouble
yum install bwa python-pip python-devel tkinter
Then:
cd /data/www/crispor
python3.9 -m venv venv
. venv/bin/activate
Get the Python packages:
pip install biopython numpy scikit-learn pandas twobitreader xlwt keras tensorflow h5py rs3 pytabix matplotlib lmdbm
(I want to remove the lmdbm dependency, wastes too many inodes.)
Keras/tensorflow is for the Cpf1 scoring model. I am using keras/tensorflow 2.1.1. I hope that the exact version is not important.
Install required R libraries if you want to use the WangSVM efficiency score (unlikely, see below):
sudo Rscript -e 'install.packages(c("e1071"), repos="http://cran.rstudio.com/")'
sudo Rscript -e 'source("https://bioconductor.org/biocLite.R"); biocLite(c("limma"));'
The R packages have not changed in many years. The version should really not matter at all. In principle, you can remove the wang score from crispor.py in the global variable where the scores are defined and not worry about R anymore. I don't think that as of 2022 anyone is still using this score for designing their guides.
When you run crispor.py, it should then show the usage message:
Usage: crispor.py [options] org fastaInFile guideOutFile
Command line interface for the Crispor tool.
org = genome identifier, like hg19 or ensHumSap
fastaInFile = Fasta file
guideOutFile = tab-sep file, one row per guide
Use "noGenome" if you only want efficiency scoring (a LOT faster). This option
will use BWA only to match the sequence to the genome, extend it and obtain
efficiency scores.
If many guides have to be scored in batch: Add GGG to them to make them valid
guides, separate these sequences by at least one "N" character and supply as a single
fasta sequence, a few dozen to ~100 per file.
Options:
-h, --help show this help message and exit
-d, --debug show debug messages, do not delete temp directory
-t, --test run internal tests
-p PAM, --pam=PAM PAM-motif to use, default NGG. TTTN triggers special
Cpf1 behavior: no scores anymore + the PAM is assumed
to be 5' of the guide. Common PAMs are:
NGG,TTTN,NGA,NGCG,NNAGAA,NGGNG,NNGRRT,NNNNGMTT,NNNNACA
-o OFFTARGETFNAME, --offtargets=OFFTARGETFNAME
write offtarget info to this filename
-m MAXOCC, --maxOcc=MAXOCC
MAXOCC parameter, guides with more matches are
excluded
--mm=MISMATCHES maximum number of mismatches, default 4
--bowtie new: use bowtie as the aligner. Do not use. Bowtie
misses many off-targets.
--skipAlign do not align the input sequence. The on-target will be
a random match with 0 mismatches.
--noEffScores do not calculate the efficiency scores
--minAltPamScore=MINALTPAMSCORE
minimum MIT off-target score for alternative PAMs, default
1.0
--worker Run as worker process: watches job queue and runs jobs
--clear clear the worker job table and exit
-g GENOMEDIR, --genomeDir=GENOMEDIR
directory with genomes, default ./genomes
To test the program, first make sure that there is a directory "../genomes". If it's not there, rename "genomes.sample" to "genomes":
mv ../genomes.sample ../genomes
Then run this command:
mkdir -p sampleFiles/mine/
crispor.py sacCer3 sampleFiles/in/sample.sacCer3.fa sampleFiles/mine/sample.sacCer3.tsv -o sampleFiles/mine/sample.sacCer3.mine.offs.tsv
The files in sampleFiles/mine should be identical to the files in sampleFiles/out/
The file testInHg19.fa contains a sample for the hg19 genome, the output is in testOutHg19.tab and testOutHg19Offtargets.tab
../crispor.py hg19 testInHg19.fa testOutHg19.mine.tab -o testOutHg19Offtargets.mine.tab
To add more genomes than yeast, skip the next section. If you want to run your script now as a web service, continue reading with the next section.
If you use a different Python version than what I use, then you will get an error message like this:
Python 3.6:
ValueError: unsupported pickle protocol: (some number)
Or:
Python 3.9 (bugfix releases)
/data/www/venv/lib/python3.9/site-packages/sklearn/base.py:318: UserWarning: Trying to unpickle estimator DecisionTreeRegressor from version 1.1.1 when using version 1.2.2. This might lead to breaking code or invalid results. Use at your own risk. For more info please refer to: https://scikit-learn.org/stable/model_persistence.html#security-maintainability-limitations
Or:
Python 3.10:
AttributeError: GradientBoostingRegressor has no attribute 'loss'
This is because these modern machine learning package authors seem to believe that they do not need to come up with data files. They always serialize their internal structures, instead of saving to a normal file. So you have to re-train the Azimuth model and save it again:
pip install openpyx # who would ever package a software with essential data stored in Excel files? You can guess who... cd bin/Azimuth-2.0/ python model_comparison.py
I kept an old version of the pickle files for python3.6 in bin/Azimuth-2.0/azimuth/saved_models/*.python3.6
Make sure you can execute CGI scripts somewhere. Your Apache config (e.g. /etc/apache2/sites-enabled/000-default) should contain a section like this:
<Directory "/var/www/html">
AllowOverride All
Options +ExecCGI (...)
AddHandler cgi-script .cgi .pl .py
Also make sure you have the CGI module enabled:
sudo a2enmod cgi
sudo service apache2 restart
If using SElinux, especially on Fedora/CentOS/RedHat, please switch it off or set it to permissive mode.
Clone the repo into such a directory:
cd /var/www/html/
git clone https://github.com/maximilianh/crisporWebsite
Use the sample E. coli genome for a start:
mv genomes.sample genomes
Create a temp directory with the right permissions:
mkdir temp
chmod a+rw temp
Make sure that Apache is allowed to execute the crispor.py script, it should have x and r permissions for all:
ls -la crispor.py
# if not ...
chmod a+rx crispor.py
By default, the jobs database is a SQlite file, /tmp/crisporJobs.db. The Apache user has to be able to write to it so let us create it now:
./crispor.py --clear
Worker queue now empty
Now start a single worker job. It will watch the job queue and process jobs:
./startWorkers.sh 1
Check that your worker is indeed running:
cat log/worker1.log
ps aux | grep crispor
Now try to access the script from a webbrowser, http://localhost/crispor.py and click "Submit"
If you want to add to your own crispor.py installation a genome that is already on crispor.org, that's very easy. All genomes available on crispor.org (except a few pre-publication ones) are provided as pre-indexed and correctly formattef files for download at http://crispor.tefor.net/genomes/. To get one of these into the current directory, use a command like this (replace hg38 with your genome code):
mkdir genomes
cd genomes
mkdir hg38
cd hg38
wget -r -l1 --no-parent -nd --reject 'index*' --reject 'robots*' http://crispor.tefor.net/genomes/hg38/
If you need to add a new genomes, this is quite a bit more involved. Ideally you want gene models in the right format (GFF), a fastsa file and various tools to convert and index these. In most cases, it's much easier to email [email protected] and ask me to add the genome, then you can download it as above. If this is not what you want, you can add a genome yourself, there even is a script for it. Look into the "tools" directory https://github.com/maximilianh/crisporWebsite/tree/master/tools, try the script crisprAddGenome. You will need to download the UCSC tools twoBitToFa
and bedToBigBed
from http://hgdownload.cse.ucsc.edu/admin/exe/linux.x86_64/ and install the tool gffread
by installing cufflinks on your machine (e.g. with apt-get install cufflinks
).
The subdirectory usrLocalBin contains other required tools for this script, you can copy them into /usr/local/bin of your machine, they are 64bit static linux binaries and should work on most current machines.
The script can auto-download genomes from Ensembl, UCSC or NCBI or allows you to add your own custom genome in .fasta format and .gff.
E.g. to add the X. laevis genome: sudo crisprAddGenome fasta /tmp2/LAEVIS_7.1.repeatMasked.fa --desc 'xenBaseLaevis71|Xenopus laevis|X. laevis|Xenbase V7.1' --gff geneModels.gff3
The four |-split values for the --desc option are: internalDatabaseName, scientificName, commonOrDisplayName, VersionNameOfAssembly
Make sure that internalDatabaseName does not include special characters, spaces etc. as it is used as a directory name.
The .bed input is always fastest, as it saves the initial BWASW step where crispor maps to the target genome.
If you are using the FASTA input, instead of feeding it a multi-fasta file (where crispor will map every piece to the genome first), try to feed it a single sequence and separate every 23bp-target in it with NN. This means that you will not get the efficiency scores but you can run these separately or in parallel with crisporEfficiencyScores.py.
For a major speedup in processing time, try to put the genome onto the ramdisk:
twoBitToFa genomes/hg19/hg19.2bit /dev/shm/hg19.fa
crispor.py will find the genome file and use bedtools to get the flanking sequences. This is almost 10x faster than the twoBitToFa command (at the cost of more RAM).
Alternatively, you may want to give flashfry by Aaron McKenna a try. It is optimized for large libraries, it uses much more RAM and has fewer scores but is sufficient for most large-library-design applications.