DOCK 3.7 2015/04/15 abl1 Tutorial: Difference between revisions

From DISI
Jump to navigation Jump to search
No edit summary
m (asdf)
 
(65 intermediate revisions by 3 users not shown)
Line 1: Line 1:


This tutoral use the 3.7.2 beta version of dock release on XXX.
This tutorial use the 3.7.2 beta version of dock release on April 17, 2015.  This tutorial is now superseded by[[DOCK 3.7 2018/06/05 abl1 Tutorial]].  


This is for a Linux environment and the scripts assume that you are running on SGE queueing system.
This is for a Linux environment and the scripts assume that you are running on SGE queuing system.
 
More information and tutorials, see [[DOCK_3.7]].  


== set up directories and get databases ==
== set up directories and get databases ==
Line 8: Line 10:
Create directory called "RotationProject"
Create directory called "RotationProject"


create a python file called "autodude_db_download.py"  
create a python file called "0000.autodude_db_download.py"  


  # this gets the database from the autodude webpage
  # this gets the database from the autodude webpage
Line 68: Line 70:
== run be_blasti.py==
== run be_blasti.py==


creat the following cshell script 0001.be_balsti_py.csh.  
First we need to get our protein of interest from the protein databank (pdb).  We will typiclly use a receptor with a ligand bound as is the case for pdbcode 2HYY, which is the Abl kinase domain in complex with imatinib (STI571, Glivec).
 
Note, in the following scripts, that DOCKBASE is a environment variable that point to the DOCK3.7 code. e.g.:
 
setenv DOCKBASE "/path2dock3.7/DOCK"
 
or
 
setenv DOCKBASE "/nfs/home/tbalius/zzz.github/DOCK"
 
Note, if you get an error ImportError: No module named Bio.PDB, then install the biopython as followed:
 
sudo yum install python-biopython
 
 
Create the following cshell script 0001.be_balsti_py.csh by using your favorite text editor (eg vim).
 
Note that the "#" symbol denotes a comment to explain what the script is doing.  


  #!/bin/csh  
  #!/bin/csh  
Line 106: Line 125:
  # the atom type definition is needed for msms which is sym-linked into the cwd
  # the atom type definition is needed for msms which is sym-linked into the cwd
   ln -s /nfs/home/tbalius/zzz.programs/msms/atmtypenumbers .
   ln -s /nfs/home/tbalius/zzz.programs/msms/atmtypenumbers .
  # carbs are disregarded as ligands! if it is: carbohydrate instead of nocarbohydrate
  # carbs are disregarded as ligands! if it is: carbohydrate instead of noncarbohydrate
  # renumber renumbers the residue number
  # renumber renumbers the residue number
   python $DOCKBASE/proteins/pdb_breaker/be_blasti.py --pdbcode $pdbname nocarbohydrate original_numbers | tee -a pdbinfo_using_biopython.log
   python $DOCKBASE/proteins/pdb_breaker/be_blasti.py --pdbcode $pdbname nocarbohydrate original_numbers | tee -a pdbinfo_using_biopython.log
Line 132: Line 151:




running 0001.be_balsti_py.csh will run a script that come with dock call be_blasti.
running 0001.be_balsti_py.csh will run a script that comes with dock called be_blasti.
And it will do the following  
 
# download the pdb file from the web,
issue the following command to run the script:
# break the file into rec and ligand componates
  csh 0001.be_balsti_py.csh
 
It will do the following  
# Download the pdb file from the web
# Break the file into rec and ligand components


Note that you will need to have msms on you system.
Note that you will need to have msms on you system.
[[get msms]]
[[get msms]]


check to make sure that the right ligand was selected and the the residue is not missing anything of importants.   
For Shoichet lab members msms is already installed.
If this automatic procigure has not perpared these files correctly then modify them.  
 
check to make sure that the right ligand was selected and the the residue is not missing anything of importance.   
If this automatic procedure has not prepared these files correctly, then modify them.  


Visulize them with chimera or an alternive visulazation program like pymol.  
Visualize them with chimera or an alternive visualization program like pymol.  


  cd 2HYY
  cd 2HYY
Line 149: Line 175:
  chimera rec.pdb lig.pdb
  chimera rec.pdb lig.pdb


[[File:rec_lig_2HYY.png|thumb|center|375px|1FJS the receptor and ligand generated from be_blasti.py.]]
[[File:rec_lig_2HYY.png|thumb|center|375px|2HYY, the receptor and ligand generated from be_blasti.py.]]


== run blastermaster.py ==
== run blastermaster.py ==


Write (paste what follows) the following script using a text editor like vi.  
Write (paste what follows) the following script using a text editor like vi. This script creates the files necessary for docking including the spheres (for orienting the ligands/decoys) and grids (for scoring the ligand/decoy poses)
 
WARNING: if you copy and pasted the script make sure there is no space before the "EOF"; this is because "EOF" designates the "end of file" for the cat command, if a space is there it wont stop cat'ing.   
 
0002.blastermaster.csh  
0002.blastermaster.csh  


Line 195: Line 224:
  #\$ -q all.q
  #\$ -q all.q
  cd $workdir
  cd $workdir
  $DOCK_BASE/src/blastermaster_1.0/blastermaster.py --addhOptions=" -HIS -FLIPs "  -v
  python $DOCKBASE/proteins/blastermaster/blastermaster.py --addhOptions=" -HIS -FLIPs "  -v
  EOF
  EOF
   
   
Line 215: Line 244:
  #    matching_spheres.sph
  #    matching_spheres.sph


== run enrichment calucaltions ==
== Visualize the docking spheres ==
Write a file called 0003.lig-decoy_enrichment.csh  
 
 
Use the showsphere program:
 
  $DOCKBASE/proteins/showsphere/bin/showsphere
 
  $DOCKBASE/proteins/showsphere/doshowsph.csh file.sph 1 file.pdb
 
for example:
$DOCKBASE/proteins/showsphere/doshowsph.csh 2HYY/dockfiles/matching_spheres.sph 1 matching_spheres.pdb
 
Alternatively, use the following simple cshell/awk script to convert spheres to pdb format:
 
cat << EOF > sphere2pdb.csh
#!/bin/csh -f
awk '$0!~/e/{ \
printf("ATOM  %5d  C  SPH%5d%12.3f%8.3f%8.3f%6.2f%6.2f\nTER\n", \$1, \$1, \$2, \$3, \$4, 1, \$5)}' \$1
EOF
 
This command statement will write the commands script to a file called sphere2pdb.csh.  Make sure there is not space before the second EOF (end of file) above.  To run the command:
 
csh sphere2pdb.csh file.sph > file.pdb
 
 
cd 2HYY/working
 
chimera rec.pdb matching_spheres.pdb
 
[[File:Spheres.png|thumb|center|375px|The spheres generated from blastermaster.py in relation to the receptor.]]
 
The box used for scoring can also be visualized in chimera with the following command:
 
chimera rec.pdb matching_spheres.pdb box
 
[[File:Spheresbox.png|thumb|center|375px|The spheres and box generated from blastermaster.py in relation to the receptor.]]
 
== run enrichment calculations ==
 
See the following tutorial for updated scripts:
[[DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_%26_Jiankun)]]
 
Here is the new recomend script that uses the preexisting submission infrastructure of DOCK:
 
[[DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_%26_Jiankun)#Submit an enrichment calculation via 0003.lig-decoy_enrichment_submit.csh]]
 
The script at the following location submits a single job as an array job, ie one script is submitted, but many task are run, where each task is a separate chunk:
 
[[DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_%26_Jiankun)# Submit an enrichment calculation via new 0003.lig-decoy_enrichment.csh]] 
The following script has advantages in that it is a very intuitive script -- every db2 chunk has a qsub script that is submitted to the queue.
 
If you decide not to use the methods linked to above then write a file called 0003.lig-decoy_enrichment.csh  
 
  #!/bin/csh
  #!/bin/csh
   
   
Line 222: Line 303:
  #It assumes that ligands and decoys have been pre-prepation (see script blablabla_ToDo) which needs to be run in SF.
  #It assumes that ligands and decoys have been pre-prepation (see script blablabla_ToDo) which needs to be run in SF.
   
   
  # filedir is where your rec.pdb and xtal-lig.pdb and dockfiles directory live  
  # $filedir/$pdbname is where your rec.pdb, xtal-lig.pdb, INDOCK, and dockfiles directory live  
  set filedir = "/mnt/nfs/home/rstein/RotationProject" #CHANGE THIS
  set filedir = "/mnt/nfs/home/rstein/RotationProject" #CHANGE THIS
  # this is where the work is done:
  # this is where the work is done:
Line 241: Line 322:
   
   
  # change if you want to use a different or consistent dock version
  # change if you want to use a different or consistent dock version
  set dock = ${DOCK_BASE}/bin/Linux/dock3.7_flex/dock.csh
  set dock = ${DOCKBASE}/docking/DOCK/bin/dock64
   
   
  set list = "1IEP"  
  set list = "2HYY"  
  #set list = `cat $1`
  #set list = `cat $1`
  #set list = `cat file`
  #set list = `cat file`
Line 316: Line 397:
   
   
  EOF
  EOF
# pause for 100 seconds if you have submitted a lot of jobs
while ( `qstat | wc -l ` > 400 )
  echo "sleep . . ."
  sleep 100
end
   
   
  qsub DOCKING_${db_type}.csh
  qsub DOCKING_${db_type}.csh
Line 327: Line 414:
  end # db_type
  end # db_type
  end # pdbname
  end # pdbname
== combine scores and poses ==
Write this file as 0004.combineScoresAndPoses.csh. This script combines your docking runs (for ligands and decoys) and stores them in the extract_all.txt file. It also creates a .mol2 file containing all top scoring poses for each ligand/decoy.
Note that if you used the alternative (which is preferred) submission in the above step, then you will need to modify the directory structure in the below script.
From:
  ${mountdir}/${pdbname}/ligands-decoys/${db_type}/allChunksCombined
To:
  ${mountdir}/${pdbname}/${db_type}/allChunksCombined
#!/bin/csh
# This script combines the results from the ligand-decoy run 0003 (all chunks) into a combine file containing dock scores from OUTDOCK files
# Three files are produced (one for lig, decoy and both)
# and: a file which has top poses as specified (e.g. top 1000 molecules with 2 poses each); two files (for lig and for decoys)
# to remove dir
# rm -fr pdbs/3O1D/ligands-decoys/ligands/allChunksCombined/ pdbs/3O1D/ligands-decoys/decoys/allChunksCombined/ pdbs/3O1D/ligands-decoys/dockedLigDecoyCombined/
set filedir = "/mnt/nfs/home/rstein/RotationProject"
set mountdir = "/mnt/nfs/home/rstein/RotationProject"
set d37 =  $DOCKBASE/analysis/
cd $mountdir
set list = "2HYY"
#set list = `cat filename`
#set list = `cat $1`
foreach pdbname ( $list )
foreach db_type ( "ligands" "decoys" )
set workdir = ${mountdir}/${pdbname}/ligands-decoys/${db_type}/allChunksCombined
echo $pdbname
#ls -l ${mountdir}/${pdbname}/${db_type}/
mkdir -p ${workdir}
cd ${workdir}
 
# creates a file called dirlist that contains the full path of all directories with docked runs (chunks)
ls -ld ${mountdir}/${pdbname}/ligands-decoys/${db_type}/* | awk '/chunk/{print $9}' > dirlist
#ls -ld ${mountdir}/${pdbname}/ligands-decoys/${db_type}/*
# for debuging
#echo "print $db_type dirlist:"
#cat dirlist
# script extracts scores from all docking runs specified in dirlist
$d37/extract_all.py
# script gets poses for top scoring molecules and produces poses.mol2 (default name)
$d37/getposes.py -d ${mountdir}/${pdbname}/ligands-decoys/${db_type}
end # db_type
## combine decoyes and actives
set workdir =  ${mountdir}/${pdbname}/ligands-decoys/dockedLigDecoyCombined
rm -rf ${workdir}
mkdir -p ${workdir}
cd ${workdir}
cat ${mountdir}/${pdbname}/ligands-decoys/ligands/allChunksCombined/dirlist ${mountdir}/${pdbname}/ligands-decoys/decoys/allChunksCombined/dirlist > dirlist
# for debuging
#echo "print ALL dirlist"
#cat dirlist
$d37/extract_all.py
#$d37/getposes.py -d ${mountdir}/${pdbname}    # doesn't work yet; not really needed
#getposes.py -z -l 1000 -x 2 -f extract_all.sort.uniq.txt -o ligands.1000.mol2 -d /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/Enrichment/1DB1/DOCKING/ligands
end # pdbname
[[File:liganddock.png|thumb|center|375px|The pose generated from a ligand (purple) docked to Abl1 compared to imatinib (blue).]]
[[File:decoydock.png|thumb|center|375px|The pose generated from a decoy (green) docked to Abl1 compared to imatinib (blue).]]
== create AUC plot of ligands and decoys ==
Make sure that you are using a version of python that has matplotlib/numpy/scipy modules:
For Shoichet user source the following:
  source /nfs/soft/python/envs/complete/latest/env.sh
or
  source /nfs/soft/python/envs/complete/latest/env.csh
Also, if you are running the processing script remotely make sure to use X11 forwarding by including a -X in the ssh command.
eg:
  ssh gimel.ucsf.bkslab.org -X
Write a file called 0005.AUCplot_of-lig-decoys.csh.
#!/bin/csh
# This script creates a log adjusted AUC (ROC) plot with ligand vs decoy results
# need X11 forwarding enabled when running remotely (ssh sgehead -X).
set filedir = "/mnt/nfs/home/rstein/RotationProject"            #CHANGE THIS
set mountdir = "/mnt/nfs/home/rstein/RotationProject"          #CHANGE THIS
set d37 =  $DOCKBASE/analysis
set dude_dir = "/mnt/nfs/home/rstein/RotationProject/databases"        # should contain decoy.smi and ligands.smi
# ln -s /mnt/nfs/work/users/fischer/VDR/lig-decoy-db/ligands.mod.smi /mnt/nfs/work/users/fischer/VDR/lig-decoy-db/ligands.smi
# CHANGE THIS
set list = "2HYY"
#set list = `cat filename`
#set list = `cat $1`
foreach pdbname ( $list )
set workdir = ${mountdir}/${pdbname}/ROC_ligdecoy/
# This script will not work without the following line:
echo "HERE is the HAWK"
# checks that previous script 0003 has produced mol2 files
if (! ( -s $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/poses.mol2) && ! (-s $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/poses.mol2 )) then
    ls -l $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/poses.mol2
    ls -l $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/poses.mol2
    echo "skipping ${pdbname}. cannot generate ROC"
    continue
endif
rm -rf $workdir
mkdir -p $workdir
cd $workdir
#wget http://dude.docking.org/targets/aa2ar/actives_final.ism
# reads ZINC ids (ligand or decoy molecule names)
# everything
awk '{print $2}' $dude_dir/decoys.ism > decoys.name  # note that you may have to change the column ($2) based on where the SMILES codes are
awk '{printf "%9s\n", $3}' $dude_dir/ligands.ism > ligands.name # note that you may have to change the column ($3) based on where the SMILES codes are
#things that finished docking
awk '{print $3}' $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/extract_all.sort.uniq.txt > decoys.finished.name
awk '{print $3}' $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/extract_all.sort.uniq.txt > ligands.finished.name
cat ${mountdir}/${pdbname}/ligands-decoys/ligands/allChunksCombined/dirlist ${mountdir}/${pdbname}/ligands-decoys/decoys/allChunksCombined/dirlist > dirlist
#which enrich.py
set enrich_py = $d37/enrich.py
set plots_py = $d37/plots.py
pwd
# calculates AUCs, stores in txt file which is then plotted for finished ligands and decoys
python ${enrich_py} -i .  -o . --ligand-file=ligands.finished.name --decoy-file=decoys.finished.name
python ${plots_py} -i . -o . --ligand-file=ligands.finished.name --decoy-file=decoys.finished.name -l $pdbname
mv roc.txt    roc.finished.txt
mv roc_own.txt roc_own.finished.txt
mv roc_own.png roc_own.finished.png
#
# calculates AUCs, stores in txt file which is then plotted for all ligands and decoys
# - i is the flag for the input directory, this dir should contain the extract_all.sort.uniq.txt.
#  the scripts enrich_py and plots_py will go through the extract file and look for the ligand and decoy names.
#  when it finds them it will populate the ROC cruve. these values are devied by the total number of ligand or decoys.
#  note that often not all ligands and not all decoys finish so the point (1,1) is always included and interpolations is performed . . .
#
#python ${enrich_py} -i $mountdir/${pdbname}/ligands-decoys/dockedLigDecoyCombined/ -o . --ligand-file=ligands.name --decoy-file=decoys.name
#python ${plots_py} -i $mountdir/${pdbname}/ligands-decoys/dockedLigDecoyCombined/ -o . --ligand-file=ligands.name --decoy-file=decoys.name -l $pdbname
python ${enrich_py} -i . -o . --ligand-file=ligands.name --decoy-file=decoys.name
python ${plots_py} -i . -o . --ligand-file=ligands.name --decoy-file=decoys.name -l $pdbname
end  #pbdname
The ROC plot for ligands and decoys docking to Abl1 looks like this:
[[File:AblROCplot.png|thumb|center|375px|The ROC plot generated for the ligands and decoys.]]
This shows poor enrichment, though it is better than random (indicated by the dotted line). This is to be expected as docking to kinases is notoriously difficult. Two changes to our procedure would be to delete some of the spheres to focus docking on a particular region of the protein we are most interested in, as well as increasing the polarity of the hinge region of Abl1 to allow for more hydrogen bonding.
== improving docking ==
===Modify matching spheres===
The docking could be potentially improved by modifying the spheres and orienting the ligands/decoys to only those spheres within the hinge region of Abl1. The following image show modified spheres:
[[File:modspheres.png|thumb|center|375px|Modified spheres to improve docking.]]
To modify spheres: first, convert them to pdb format; then, visualize them in your favorite program (pymol, chimera, etc.); then, delete, move, or add atoms to this file; and finally convert it back to the sph format. 
As discussed above [[http://wiki.docking.org/index.php/DOCK_3.7_2015/04/15_abl1_Tutorial#Visualize_the_docking_spheres]], you may use doshowsph.csh to convert the spheres to pdb format.
Here is the program that will convert a pdb file into a sphere file (this program take 2 inputs: name of pdbfile to convert and the name of the sphere file that you want to create.).
$DOCKBASE/proteins/pdbtosph/bin/pdbtosph
for example:
$DOCKBASE/proteins/pdbtosph/bin/pdbtosph matching_spheres_mod.pdb matching_spheres_mod.sph
The 0003.lig-decoy_enrichment.csh would be modified to include this line and renamed to 0003.lig-decoy_enrichment_mod_sph.csh:
!/bin/csh
#This script docks a DUD-e like ligand-decoy-database to evaluate the enrichment performance of actives over decoys
#It assumes that ligands and decoys have been pre-prepation (see script blablabla_ToDo) which needs to be run in SF.
# filedir is where your rec.pdb and xtal-lig.pdb and dockfiles directory live
set filedir = "/mnt/nfs/home/rstein/RotationProject"    #CHANGE THIS
# this is where the work is done:
set mountdir = $filedir                        # Might CHANGE THIS
set dude_dir = "/mnt/nfs/home/rstein/RotationProject/databases"  # should contain decoy.smi and ligand.smi for ROC script 00005...csh
  ## TO DO - rename this outside in the dir structure and call in blbalbalbabla script
if (-s $dude_dir) then
  echo " $dude_dir exist"
else
  # this is something to modified in future.
  # probably better to exit if it is not there.
  echo "databases do not exist. "
  echo "consider making a symbolic link to the database files"
  #echo "making a symbolic link:"
#echo "ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir"
  #ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir
endif
# change if you want to use a different or consistent dock version
set dock = ${DOCKBASE}/docking/DOCK/bin/dock64
set list = "2HYY"
#set list = `cat $1`
#set list = `cat file`
                                # CHANGE THIS (pdbname)
foreach pdbname ( $list )
# creates "ligands" and "decoys" and has the aim to dock all of the subsets for those two
foreach db_type ( "ligands" "decoys" )
set workdir1 = "${mountdir}/${pdbname}/ligands-decoys_sphmod/${db_type}"
mkdir -p  ${workdir1}
cd  ${workdir1}
# puts dockfiles in the right relative-path that INDOCK file expects
ln -s $filedir/${pdbname}/dockfiles .
set count = '1'
# loop over database files to put each into a seperate chunk
foreach dbfile (`ls $dude_dir/${db_type}/${db_type}*.db2.gz`)
echo $dbfile
set chunk = "chunk$count"
set workdir2 = ${workdir1}/$chunk
## so you don't blow away stuff
if ( -s $workdir2 ) then
    echo "$workdir2 exits"
    continue
endif
#rm -rf ${workdir}
mkdir -p ${workdir2}
cd ${workdir2}
# copy INDOCK file of choice in right location
#cp $filedir/zzz.dock3_input/INDOCK .
#cp $filedir/INDOCK_match20K INDOCK
#cp $filedir/INDOCK_5k_TolerantClash INDOCK    # CHANGE THIS
cp $filedir/${pdbname}/INDOCK .
# modified the dock file using sed. here we change some key sampling parameters; sed -i changes input file internally (overwrites), -e changes file externally (pipes it to screen or into file if redirected)
#sed -i "s/bump_maximum                  50.0/bump_maximum                  500.0/g" INDOCK
#sed -i "s/bump_rigid                    50.0/bump_rigid                    500.0/g" INDOCK
#sed -i "s/check_clashes                yes/check_clashes                no/g" INDOCK
sed -i "s/receptor_sphere_file          ..\/dockfiles\/matching_spheres.sph/receptor_sphere_file          ..\/..\/..\/working\/matching_spheres_mod.sph/g" INDOCK
ln -s $dbfile .
set dbf = `ls *.gz`
echo "./$dbf"
# says what to dock and where it sits
echo "./$dbf" > split_database_index\
# writes submission script that runs dock on the sgehead queue
cat <<EOF > DOCKING_${db_type}.csh
#\$ -S /bin/csh
#\$ -cwd
#\$ -q all.q
#\$ -o stdout
#\$ -e stderr
cd ${workdir2}
echo "starting . . ."
date
echo $dock
$dock
date
echo "finished . . ."
EOF
qsub DOCKING_${db_type}.csh
# alternatively if you don't want to run it on the queue but locally comment in this instead:
#csh DOCKING_${lig_type}.csh &
@ count = ${count} + 1
# counter is chuch dir
end # dbfile
end # db_type
end # pdbname
=== make the hing region more polar ===
see the following page:
[[DOCK_3.7_tart]]
==Virtual Screening==
===database setup===
This part of the tutorial is tailored for shoichet lab use.  An outside user of dock might need to deviate from what is described.
Go to zinc and select your compounds of interested:
'[http://zinc15.docking.org/tranches/home http://zinc15.docking.org/tranches/home]'
This is the tranches page which allows users to select the region of chemical space of interest.
lets select the fragment preset.
on lets download the index file.  This file contains the location of each database on our cluster.  outside users will need to download the databases themselves.   
now lets setup the directorys for docking by running the following script:
python /nfs/home/tbalius/zzz.github/DOCK/docking/setup/setup_db2_zinc15_file_number.py ./ vs_frag /nfs/work/tbalius/database_ph4/frags.txt 500  count
The file /nfs/work/tbalius/database_ph4/frags.txt should be changed to that you downloaded from ZINC. 
The above script has 5 parameters:
:(1) path where directories will be located (present directory);
:(2) prefix name of the directories;
:(3) the file that contains the db2 files locations;
:(4) the number of directories to be created; and
:(5) the type of run: count (evenly distributes the db2 file among the dirs, this is much faster than the other options), size (It will try and make the directory of equal size), or both (will try and satisfy both criteria). 
Note that this script is avable in later beta versons of DOCK3.7.
===submitting the docking calculations===
This script will submit a job to the queue for each of the docking directorys created by the setup script.
$DOCKBASE/docking/submit/submit.csh
DOCK3.7 is a serial program and is parallelized by submiting many serial jobs to the queue.
===combining the results===
After your docking jobs have all completed, This script will combine all your results into an extract_all file.
  $DOCKBASE/analysis/extract_all.py
This script will create a mol2 file with the top scoring molecules:
  $DOCKBASE/analysis/getposes.py
=== curating and hit-picking ===
Typically we will visualize the ligands in UCSF Chimera using the veiwdock tool (using the DOCK4, 5, or 6 format).

Latest revision as of 21:05, 19 October 2023

This tutorial use the 3.7.2 beta version of dock release on April 17, 2015. This tutorial is now superseded byDOCK 3.7 2018/06/05 abl1 Tutorial.

This is for a Linux environment and the scripts assume that you are running on SGE queuing system.

More information and tutorials, see DOCK_3.7.

set up directories and get databases

Create directory called "RotationProject"

create a python file called "0000.autodude_db_download.py"

# this gets the database from the autodude webpage

import sys, os
import urllib

system = 'abl1'
url = 'http://autodude.docking.org/dude_e_db2/'

print "url = " + url

#page=requests.get(url)

webfile = urllib.urlopen(url)
page    = webfile.read()
webfile.close()

splitpage=page.split('\n')

for line in splitpage:
   if system in line: 
      file = line.replace('"',' ').split()[2]
      print url+file
      urllib.urlretrieve(url+file,file)

     # exit()

This python script will download the dockable db2 databases from the autodude webpage.

python /mnt/nfs/home/rstein/RotationProject/autodude_db_download.py 

make a subdirectory called databases:

mkdir databases

go inside.

cd databases

make directories for ligands and decoys and move the corresponding files into those directories

mkdir decoys 
mv decoys*db2.gz decoys
mkdir ligands 
mv ligands*db2.gz ligands

download the ligand and decoy isomeric smiles file:

wget http://autodude.docking.org/abl1/decoys_final.ism
mv decoys_final.ism decoys.ism

note that the scripts expect the name to be decoys.ism, so we changed the name.

wget http://autodude.docking.org/abl1/actives_final.ism
mv actives_final.ism ligands.ism

run be_blasti.py

First we need to get our protein of interest from the protein databank (pdb). We will typiclly use a receptor with a ligand bound as is the case for pdbcode 2HYY, which is the Abl kinase domain in complex with imatinib (STI571, Glivec).

Note, in the following scripts, that DOCKBASE is a environment variable that point to the DOCK3.7 code. e.g.:

setenv DOCKBASE "/path2dock3.7/DOCK"

or

setenv DOCKBASE "/nfs/home/tbalius/zzz.github/DOCK"

Note, if you get an error ImportError: No module named Bio.PDB, then install the biopython as followed:

sudo yum install python-biopython


Create the following cshell script 0001.be_balsti_py.csh by using your favorite text editor (eg vim).

Note that the "#" symbol denotes a comment to explain what the script is doing.

#!/bin/csh 

# this script calls be_blasti.py which creates a receptor and ligand file from a (list of) pdbcode(s).

# msms is a molecular surface generation program needed for be_blasti.py to run
# which is put in your path
set path = ( /nfs/home/tbalius/zzz.programs/msms $path )
# you will need to have msms on you system.   

set list = "2HYY" # or use `cat filename` to list your pdb codes here from a text file like pdblist_rat, to loop over each variable (pdb code) later
#set list = `cat $1`
#set list = `cat /nfs/work/users/tbalius/VDR/Enrichment/pdblist_rat `

# CHANGE THIS, according to where the magic is going to happen
#set mountdir = "/mnt/nfs/work/users/tbalius/VDR/"
set mountdir = `pwd` 

# loop over pdbnames e.g. 1DB1 or list
foreach pdbname ( $list )

echo " ${pdbname} "

# for each pdb makes a directory with its name
set workdir = ${mountdir}/${pdbname}

## so you don't blow away stuff; continue means STOP here and continue with next pdb from list
if ( -s $workdir ) then
   echo "$workdir exits"
   continue
endif

  mkdir -p ${workdir}
  cd ${workdir}

# the atom type definition is needed for msms which is sym-linked into the cwd
  ln -s /nfs/home/tbalius/zzz.programs/msms/atmtypenumbers .
# carbs are disregarded as ligands! if it is: carbohydrate instead of noncarbohydrate
# renumber renumbers the residue number
  python $DOCKBASE/proteins/pdb_breaker/be_blasti.py --pdbcode $pdbname nocarbohydrate original_numbers | tee -a pdbinfo_using_biopython.log

# error checking looks for receptor and ligand file which should be produced by be_blasti.py
  if !(-s rec.pdb) then
      echo "rec.pdb is not found"
  endif

  mv rec.pdb temp.pdb
  grep -v TER temp.pdb | grep -v END  > rec.pdb

  rm temp.pdb

# be_blasti.py produces peptide which may be used as a ligand if no other ligand is produced
  if (-s lig.pdb) then
     sed -e "s/HETATM/ATOM  /g" lig.pdb > xtal-lig.pdb
  else if (-s pep.pdb) then ## if no ligand and peptide
     sed -e "s/HETATM/ATOM  /g" pep.pdb > xtal-lig.pdb
  else
     echo "Warning: No ligand or peptid."
  endif

end # system


running 0001.be_balsti_py.csh will run a script that comes with dock called be_blasti.

issue the following command to run the script:

 csh 0001.be_balsti_py.csh
 

It will do the following

  1. Download the pdb file from the web
  2. Break the file into rec and ligand components

Note that you will need to have msms on you system.

get msms

For Shoichet lab members msms is already installed.

check to make sure that the right ligand was selected and the the residue is not missing anything of importance. If this automatic procedure has not prepared these files correctly, then modify them.

Visualize them with chimera or an alternive visualization program like pymol.

cd 2HYY
chimera rec.pdb lig.pdb
2HYY, the receptor and ligand generated from be_blasti.py.

run blastermaster.py

Write (paste what follows) the following script using a text editor like vi. This script creates the files necessary for docking including the spheres (for orienting the ligands/decoys) and grids (for scoring the ligand/decoy poses)

WARNING: if you copy and pasted the script make sure there is no space before the "EOF"; this is because "EOF" designates the "end of file" for the cat command, if a space is there it wont stop cat'ing.

0002.blastermaster.csh

#!/bin/csh 

# This script runs Ryan's blastermaster python masterscript for generating everything that dock needs, i.e. grids, spheres
# Run on sgehead as jobs are submitted to the queue

# list is same as in 001... script 
set list = "2HYY"
#set list = `cat $1`
#set list = `cat /nfs/work/users/tbalius/VDR/Enrichment/pdblist_all `

set mountdir = `pwd`
#set mountdir = "/nfs/work/users/tbalius/VDR/"

# loop over all pdb(s)
foreach pdbname ( $list )

echo "${pdbname}"

set workdir = ${mountdir}/${pdbname}

# checks that 001 ran successfully and produced the directory structure as expected
# if not stops with current pdb code and continues with next one in list
  if ! ( -s $workdir ) then
     echo "$workdir does not exit"
     continue
  endif

cd $workdir

#cat xtal-lig_ori.pdb | awk '{if ($1 == "ATOM" || $1 == "HETATM"){print $0}}' | sed -e "s/HETATM/ATOM  /g"  >  xtal-lig.pdb

# the following lines create a qsub script which submits blastermaster to the queue
cat <<EOF > qsub.csh
#!/bin/csh 
#\$ -cwd
#\$ -j yes
#\$ -o stderr
#\$ -q all.q
cd $workdir
python $DOCKBASE/proteins/blastermaster/blastermaster.py --addhOptions=" -HIS -FLIPs "  -v
EOF

qsub qsub.csh 

end # pdbname
# going to the next pdb

# this will produce two directories:
# 1) working - contains all input and output files that are generated; not needed afterwards but as a reference
# 2) dockfiles - contains everything that is needed to run dock (copied from working)
#    grids 
#    	trim.electrostatics.phi 
#    	vdw.vdw 
#    	vdw.bmp 
# 	ligand.desolv.heavy
# 	ligand.desolv.hydrogen
#    spheres
#    	matching_spheres.sph

Visualize the docking spheres

Use the showsphere program:

 $DOCKBASE/proteins/showsphere/bin/showsphere 
 $DOCKBASE/proteins/showsphere/doshowsph.csh file.sph 1 file.pdb

for example:

$DOCKBASE/proteins/showsphere/doshowsph.csh 2HYY/dockfiles/matching_spheres.sph 1 matching_spheres.pdb

Alternatively, use the following simple cshell/awk script to convert spheres to pdb format:

cat << EOF > sphere2pdb.csh
#!/bin/csh -f
awk '$0!~/e/{ \
printf("ATOM  %5d  C   SPH%5d%12.3f%8.3f%8.3f%6.2f%6.2f\nTER\n", \$1, \$1, \$2, \$3, \$4, 1, \$5)}' \$1
EOF

This command statement will write the commands script to a file called sphere2pdb.csh. Make sure there is not space before the second EOF (end of file) above. To run the command:

csh sphere2pdb.csh file.sph > file.pdb


cd 2HYY/working

chimera rec.pdb matching_spheres.pdb

The spheres generated from blastermaster.py in relation to the receptor.

The box used for scoring can also be visualized in chimera with the following command:

chimera rec.pdb matching_spheres.pdb box

The spheres and box generated from blastermaster.py in relation to the receptor.

run enrichment calculations

See the following tutorial for updated scripts: DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_&_Jiankun)

Here is the new recomend script that uses the preexisting submission infrastructure of DOCK:

DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_&_Jiankun)#Submit an enrichment calculation via 0003.lig-decoy_enrichment_submit.csh

The script at the following location submits a single job as an array job, ie one script is submitted, but many task are run, where each task is a separate chunk:

DOCK_3.7_2016/09/16_Tutorial_for_Enrichment_Calculations_(Trent_&_Jiankun)# Submit an enrichment calculation via new 0003.lig-decoy_enrichment.csh

The following script has advantages in that it is a very intuitive script -- every db2 chunk has a qsub script that is submitted to the queue.

If you decide not to use the methods linked to above then write a file called 0003.lig-decoy_enrichment.csh

#!/bin/csh

#This script docks a DUD-e like ligand-decoy-database to evaluate the enrichment performance of actives over decoys
#It assumes that ligands and decoys have been pre-prepation (see script blablabla_ToDo) which needs to be run in SF.

# $filedir/$pdbname is where your rec.pdb, xtal-lig.pdb, INDOCK, and dockfiles directory live 
set filedir = "/mnt/nfs/home/rstein/RotationProject"	#CHANGE THIS
# this is where the work is done:
set mountdir = $filedir				# Might CHANGE THIS
set dude_dir = "/mnt/nfs/home/rstein/RotationProject/databases"  # should contain decoy.smi and ligand.smi for ROC script 00005...csh
  ## TO DO - rename this outside in the dir structure and call in blbalbalbabla script
if (-s $dude_dir) then 
  echo " $dude_dir exist"
else
  # this is something to modified in future. 
  # probably better to exit if it is not there.
  echo "databases do not exist. "
  echo "consider making a symbolic link to the database files"
  #echo "making a symbolic link:"
  #echo "ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir"
  #ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir
endif

# change if you want to use a different or consistent dock version
set dock = ${DOCKBASE}/docking/DOCK/bin/dock64

set list = "2HYY" 
#set list = `cat $1`
#set list = `cat file`
        			# CHANGE THIS (pdbname)
foreach pdbname ( $list )

# creates "ligands" and "decoys" and has the aim to dock all of the subsets for those two
foreach db_type ( "ligands" "decoys" )

set workdir1 = "${mountdir}/${pdbname}/ligands-decoys/${db_type}"

mkdir -p  ${workdir1}
cd  ${workdir1} 
# puts dockfiles in the right relative-path that INDOCK file expects
ln -s $filedir/${pdbname}/dockfiles .

set count = '1'

# loop over database files to put each into a seperate chunk
foreach dbfile (`ls $dude_dir/${db_type}/${db_type}*.db2.gz`)

echo $dbfile

set chunk = "chunk$count"

set workdir2 = ${workdir1}/$chunk

## so you don't blow away stuff
if ( -s $workdir2 ) then
   echo "$workdir2 exits"
   continue
endif

#rm -rf ${workdir}
mkdir -p ${workdir2}
cd ${workdir2}

# copy INDOCK file of choice in right location
#cp $filedir/zzz.dock3_input/INDOCK . 
#cp $filedir/INDOCK_match20K INDOCK
#cp $filedir/INDOCK_5k_TolerantClash INDOCK	# CHANGE THIS
cp $filedir/${pdbname}/INDOCK .
 # modified the dock file using sed. here we change some key sampling parameters; sed -i changes input file internally (overwrites), -e changes file externally (pipes it to screen or into file if redirected)
#sed -i "s/bump_maximum                  50.0/bump_maximum                  500.0/g" INDOCK 
#sed -i "s/bump_rigid                    50.0/bump_rigid                    500.0/g" INDOCK 
#sed -i "s/check_clashes                 yes/check_clashes                 no/g" INDOCK 

ln -s $dbfile . 

set dbf = `ls *.gz`

echo "./$dbf"

# says what to dock and where it sits
echo "./$dbf" > split_database_index

# writes submission script that runs dock on the sgehead queue
cat <<EOF > DOCKING_${db_type}.csh
#\$ -S /bin/csh
#\$ -cwd
#\$ -q all.q
#\$ -o stdout
#\$ -e stderr

cd ${workdir2}
echo "starting . . ."
date
echo $dock 
$dock
date
echo "finished . . ."

EOF

# pause for 100 seconds if you have submitted a lot of jobs
while ( `qstat | wc -l ` > 400 )
  echo "sleep . . ."
  sleep 100
end

qsub DOCKING_${db_type}.csh
# alternatively if you don't want to run it on the queue but locally comment in this instead:
#csh DOCKING_${lig_type}.csh &

@ count = ${count} + 1 
# counter is chuch dir

end # dbfile
end # db_type
end # pdbname

combine scores and poses

Write this file as 0004.combineScoresAndPoses.csh. This script combines your docking runs (for ligands and decoys) and stores them in the extract_all.txt file. It also creates a .mol2 file containing all top scoring poses for each ligand/decoy.

Note that if you used the alternative (which is preferred) submission in the above step, then you will need to modify the directory structure in the below script.

From:

 ${mountdir}/${pdbname}/ligands-decoys/${db_type}/allChunksCombined

To:

 ${mountdir}/${pdbname}/${db_type}/allChunksCombined
#!/bin/csh

# This script combines the results from the ligand-decoy run 0003 (all chunks) into a combine file containing dock scores from OUTDOCK files
# Three files are produced (one for lig, decoy and both) 
# and: a file which has top poses as specified (e.g. top 1000 molecules with 2 poses each); two files (for lig and for decoys)


# to remove dir
# rm -fr pdbs/3O1D/ligands-decoys/ligands/allChunksCombined/ pdbs/3O1D/ligands-decoys/decoys/allChunksCombined/ pdbs/3O1D/ligands-decoys/dockedLigDecoyCombined/

set filedir = "/mnt/nfs/home/rstein/RotationProject"
set mountdir = "/mnt/nfs/home/rstein/RotationProject"
set d37 =  $DOCKBASE/analysis/

cd $mountdir

set list = "2HYY"
#set list = `cat filename`
#set list = `cat $1`

foreach pdbname ( $list )

foreach db_type ( "ligands" "decoys" )

set workdir = ${mountdir}/${pdbname}/ligands-decoys/${db_type}/allChunksCombined

echo $pdbname

#ls -l ${mountdir}/${pdbname}/${db_type}/

mkdir -p ${workdir}
cd ${workdir}
 
# creates a file called dirlist that contains the full path of all directories with docked runs (chunks)
ls -ld ${mountdir}/${pdbname}/ligands-decoys/${db_type}/* | awk '/chunk/{print $9}' > dirlist

#ls -ld ${mountdir}/${pdbname}/ligands-decoys/${db_type}/*

# for debuging
#echo "print $db_type dirlist:"
#cat dirlist

# script extracts scores from all docking runs specified in dirlist
$d37/extract_all.py
# script gets poses for top scoring molecules and produces poses.mol2 (default name)
$d37/getposes.py -d ${mountdir}/${pdbname}/ligands-decoys/${db_type}

end # db_type

## combine decoyes and actives
set workdir =  ${mountdir}/${pdbname}/ligands-decoys/dockedLigDecoyCombined

rm -rf ${workdir}
mkdir -p ${workdir}
cd ${workdir}

cat ${mountdir}/${pdbname}/ligands-decoys/ligands/allChunksCombined/dirlist ${mountdir}/${pdbname}/ligands-decoys/decoys/allChunksCombined/dirlist > dirlist

# for debuging
#echo "print ALL dirlist"
#cat dirlist

$d37/extract_all.py
#$d37/getposes.py -d ${mountdir}/${pdbname}     # doesn't work yet; not really needed
#getposes.py -z -l 1000 -x 2 -f extract_all.sort.uniq.txt -o ligands.1000.mol2 -d /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/Enrichment/1DB1/DOCKING/ligands

end # pdbname


The pose generated from a ligand (purple) docked to Abl1 compared to imatinib (blue).


The pose generated from a decoy (green) docked to Abl1 compared to imatinib (blue).

create AUC plot of ligands and decoys

Make sure that you are using a version of python that has matplotlib/numpy/scipy modules:

For Shoichet user source the following:

 source /nfs/soft/python/envs/complete/latest/env.sh

or

 source /nfs/soft/python/envs/complete/latest/env.csh

Also, if you are running the processing script remotely make sure to use X11 forwarding by including a -X in the ssh command. eg:

 ssh gimel.ucsf.bkslab.org -X 


Write a file called 0005.AUCplot_of-lig-decoys.csh.

#!/bin/csh

# This script creates a log adjusted AUC (ROC) plot with ligand vs decoy results
# need X11 forwarding enabled when running remotely (ssh sgehead -X). 

set filedir = "/mnt/nfs/home/rstein/RotationProject"            #CHANGE THIS
set mountdir = "/mnt/nfs/home/rstein/RotationProject"           #CHANGE THIS
set d37 =  $DOCKBASE/analysis
set dude_dir = "/mnt/nfs/home/rstein/RotationProject/databases"         # should contain decoy.smi and ligands.smi


# ln -s /mnt/nfs/work/users/fischer/VDR/lig-decoy-db/ligands.mod.smi /mnt/nfs/work/users/fischer/VDR/lig-decoy-db/ligands.smi

# CHANGE THIS
set list = "2HYY"
#set list = `cat filename`
#set list = `cat $1`

foreach pdbname ( $list )

set workdir = ${mountdir}/${pdbname}/ROC_ligdecoy/

# This script will not work without the following line:
echo "HERE is the HAWK"

# checks that previous script 0003 has produced mol2 files
if (! ( -s $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/poses.mol2) && ! (-s $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/poses.mol2 )) then
   ls -l $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/poses.mol2
   ls -l $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/poses.mol2
   echo "skipping ${pdbname}. cannot generate ROC"
   continue
endif


rm -rf $workdir
mkdir -p $workdir
cd $workdir

#wget http://dude.docking.org/targets/aa2ar/actives_final.ism

# reads ZINC ids (ligand or decoy molecule names)
# everything
awk '{print $2}' $dude_dir/decoys.ism > decoys.name   # note that you may have to change the column ($2) based on where the SMILES codes are
awk '{printf "%9s\n", $3}' $dude_dir/ligands.ism > ligands.name # note that you may have to change the column ($3) based on where the SMILES codes are
#things that finished docking
awk '{print $3}' $mountdir/${pdbname}/ligands-decoys/decoys/allChunksCombined/extract_all.sort.uniq.txt > decoys.finished.name
awk '{print $3}' $mountdir/${pdbname}/ligands-decoys/ligands/allChunksCombined/extract_all.sort.uniq.txt > ligands.finished.name

cat ${mountdir}/${pdbname}/ligands-decoys/ligands/allChunksCombined/dirlist ${mountdir}/${pdbname}/ligands-decoys/decoys/allChunksCombined/dirlist > dirlist

#which enrich.py
set enrich_py = $d37/enrich.py
set plots_py = $d37/plots.py

pwd
# calculates AUCs, stores in txt file which is then plotted for finished ligands and decoys
python ${enrich_py} -i .  -o . --ligand-file=ligands.finished.name --decoy-file=decoys.finished.name
python ${plots_py} -i . -o . --ligand-file=ligands.finished.name --decoy-file=decoys.finished.name -l $pdbname

mv roc.txt     roc.finished.txt
mv roc_own.txt roc_own.finished.txt
mv roc_own.png roc_own.finished.png

# 
# calculates AUCs, stores in txt file which is then plotted for all ligands and decoys
# - i is the flag for the input directory, this dir should contain the extract_all.sort.uniq.txt.
#  the scripts enrich_py and plots_py will go through the extract file and look for the ligand and decoy names.
#  when it finds them it will populate the ROC cruve. these values are devied by the total number of ligand or decoys.
#  note that often not all ligands and not all decoys finish so the point (1,1) is always included and interpolations is performed . . . 
#
#python ${enrich_py} -i $mountdir/${pdbname}/ligands-decoys/dockedLigDecoyCombined/ -o . --ligand-file=ligands.name --decoy-file=decoys.name 
#python ${plots_py} -i $mountdir/${pdbname}/ligands-decoys/dockedLigDecoyCombined/ -o . --ligand-file=ligands.name --decoy-file=decoys.name -l $pdbname 
python ${enrich_py} -i . -o . --ligand-file=ligands.name --decoy-file=decoys.name
python ${plots_py} -i . -o . --ligand-file=ligands.name --decoy-file=decoys.name -l $pdbname 

end   #pbdname


The ROC plot for ligands and decoys docking to Abl1 looks like this:

The ROC plot generated for the ligands and decoys.

This shows poor enrichment, though it is better than random (indicated by the dotted line). This is to be expected as docking to kinases is notoriously difficult. Two changes to our procedure would be to delete some of the spheres to focus docking on a particular region of the protein we are most interested in, as well as increasing the polarity of the hinge region of Abl1 to allow for more hydrogen bonding.

improving docking

Modify matching spheres

The docking could be potentially improved by modifying the spheres and orienting the ligands/decoys to only those spheres within the hinge region of Abl1. The following image show modified spheres:

Modified spheres to improve docking.

To modify spheres: first, convert them to pdb format; then, visualize them in your favorite program (pymol, chimera, etc.); then, delete, move, or add atoms to this file; and finally convert it back to the sph format.

As discussed above [[1]], you may use doshowsph.csh to convert the spheres to pdb format.

Here is the program that will convert a pdb file into a sphere file (this program take 2 inputs: name of pdbfile to convert and the name of the sphere file that you want to create.).

$DOCKBASE/proteins/pdbtosph/bin/pdbtosph

for example:

$DOCKBASE/proteins/pdbtosph/bin/pdbtosph matching_spheres_mod.pdb matching_spheres_mod.sph


The 0003.lig-decoy_enrichment.csh would be modified to include this line and renamed to 0003.lig-decoy_enrichment_mod_sph.csh:

!/bin/csh

#This script docks a DUD-e like ligand-decoy-database to evaluate the enrichment performance of actives over decoys
#It assumes that ligands and decoys have been pre-prepation (see script blablabla_ToDo) which needs to be run in SF.

# filedir is where your rec.pdb and xtal-lig.pdb and dockfiles directory live 
set filedir = "/mnt/nfs/home/rstein/RotationProject"    #CHANGE THIS
# this is where the work is done:
set mountdir = $filedir                         # Might CHANGE THIS
set dude_dir = "/mnt/nfs/home/rstein/RotationProject/databases"  # should contain decoy.smi and ligand.smi for ROC script 00005...csh
  ## TO DO - rename this outside in the dir structure and call in blbalbalbabla script
if (-s $dude_dir) then
  echo " $dude_dir exist"
else
  # this is something to modified in future. 
  # probably better to exit if it is not there.
  echo "databases do not exist. "
  echo "consider making a symbolic link to the database files"
  #echo "making a symbolic link:"
#echo "ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir"
  #ln -s /mnt/nfs/work/users/fischer/VDR/27Jan2014_learningDOCKrgc/databases_all_xtal-ligand_decoy $dude_dir
endif

# change if you want to use a different or consistent dock version
set dock = ${DOCKBASE}/docking/DOCK/bin/dock64

set list = "2HYY"
#set list = `cat $1`
#set list = `cat file`
                                # CHANGE THIS (pdbname)
foreach pdbname ( $list )

# creates "ligands" and "decoys" and has the aim to dock all of the subsets for those two
foreach db_type ( "ligands" "decoys" )
set workdir1 = "${mountdir}/${pdbname}/ligands-decoys_sphmod/${db_type}"

mkdir -p  ${workdir1}
cd  ${workdir1}
# puts dockfiles in the right relative-path that INDOCK file expects
ln -s $filedir/${pdbname}/dockfiles .

set count = '1'

# loop over database files to put each into a seperate chunk
foreach dbfile (`ls $dude_dir/${db_type}/${db_type}*.db2.gz`)

echo $dbfile

set chunk = "chunk$count"

set workdir2 = ${workdir1}/$chunk
## so you don't blow away stuff
if ( -s $workdir2 ) then
   echo "$workdir2 exits"
   continue
endif

#rm -rf ${workdir}
mkdir -p ${workdir2}
cd ${workdir2}

# copy INDOCK file of choice in right location
#cp $filedir/zzz.dock3_input/INDOCK . 
#cp $filedir/INDOCK_match20K INDOCK
#cp $filedir/INDOCK_5k_TolerantClash INDOCK     # CHANGE THIS
cp $filedir/${pdbname}/INDOCK .
# modified the dock file using sed. here we change some key sampling parameters; sed -i changes input file internally (overwrites), -e changes file externally (pipes it to screen or into file if redirected)
#sed -i "s/bump_maximum                  50.0/bump_maximum                  500.0/g" INDOCK 
#sed -i "s/bump_rigid                    50.0/bump_rigid                    500.0/g" INDOCK 
#sed -i "s/check_clashes                 yes/check_clashes                 no/g" INDOCK 
sed -i "s/receptor_sphere_file          ..\/dockfiles\/matching_spheres.sph/receptor_sphere_file          ..\/..\/..\/working\/matching_spheres_mod.sph/g" INDOCK

ln -s $dbfile .

set dbf = `ls *.gz`

echo "./$dbf"

# says what to dock and where it sits
echo "./$dbf" > split_database_index\
# writes submission script that runs dock on the sgehead queue
cat <<EOF > DOCKING_${db_type}.csh
#\$ -S /bin/csh
#\$ -cwd
#\$ -q all.q
#\$ -o stdout
#\$ -e stderr

cd ${workdir2}
echo "starting . . ."
date
echo $dock 
$dock
date
echo "finished . . ."

EOF
qsub DOCKING_${db_type}.csh
# alternatively if you don't want to run it on the queue but locally comment in this instead:
#csh DOCKING_${lig_type}.csh &

@ count = ${count} + 1
# counter is chuch dir

end # dbfile
end # db_type
end # pdbname

make the hing region more polar

see the following page: DOCK_3.7_tart

Virtual Screening

database setup

This part of the tutorial is tailored for shoichet lab use. An outside user of dock might need to deviate from what is described.

Go to zinc and select your compounds of interested:

'http://zinc15.docking.org/tranches/home'

This is the tranches page which allows users to select the region of chemical space of interest.

lets select the fragment preset.

on lets download the index file. This file contains the location of each database on our cluster. outside users will need to download the databases themselves.

now lets setup the directorys for docking by running the following script:

python /nfs/home/tbalius/zzz.github/DOCK/docking/setup/setup_db2_zinc15_file_number.py ./ vs_frag /nfs/work/tbalius/database_ph4/frags.txt 500  count

The file /nfs/work/tbalius/database_ph4/frags.txt should be changed to that you downloaded from ZINC.

The above script has 5 parameters:

(1) path where directories will be located (present directory);
(2) prefix name of the directories;
(3) the file that contains the db2 files locations;
(4) the number of directories to be created; and
(5) the type of run: count (evenly distributes the db2 file among the dirs, this is much faster than the other options), size (It will try and make the directory of equal size), or both (will try and satisfy both criteria).

Note that this script is avable in later beta versons of DOCK3.7.

submitting the docking calculations

This script will submit a job to the queue for each of the docking directorys created by the setup script.

$DOCKBASE/docking/submit/submit.csh

DOCK3.7 is a serial program and is parallelized by submiting many serial jobs to the queue.

combining the results

After your docking jobs have all completed, This script will combine all your results into an extract_all file.

 $DOCKBASE/analysis/extract_all.py

This script will create a mol2 file with the top scoring molecules:

 $DOCKBASE/analysis/getposes.py

curating and hit-picking

Typically we will visualize the ligands in UCSF Chimera using the veiwdock tool (using the DOCK4, 5, or 6 format).