Skip to content

Pipeline for eukaryotic genome annotation based on external evidences

Notifications You must be signed in to change notification settings

galikbence/genome_annotation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 

Repository files navigation

GENOME ANNOTATION WORKFLOW

Workflow for eukaryotic genome annotation based on external evidences using AUGUSTUS and GeneMark-ES gene prediction tools.

workflow

Figure. Genome annotation workflow.

The workflow contains the following steps:

  1. Collecting external evidences
  2. Preparing gene models
  3. Repeat modelling & masking
  4. Predicting tRNAs
  5. Predicting protein coding genes
  6. Combining gene models
  7. Annotating protein coding genes

The following tools/packages/databases are mentioned/used in the workflow:

1. Collecting external evidences

We assume you already sequenced and assembled your genome.

In order to maximize the genome annotation efficiency you should collect such data that can support gene models. For example protein and transcript sequences from closely related species. It is highly recommended to download these datasets from reliable source!!! The best if you have RNA-seq data directly from your species of interest. You can assemble the transcripts using the Trinity transcriptome assembly tool (either applying the genome-guided method for better results) and annotate them running the Trinotate pipeline.

In the next steps we will show how to use the collected/generated data to build gene models for more accurate gene prediction.

In this workflow, we mainly focus on how to integrate data generated with short read sequencing platforms. Also, you can include evidences, specially whole length transcripts, that were generated using long read sequencing technologies (PacBio, MinION).

2. Preparing gene models and other evidences

In this section we will train AUGUSTUS for another species by the following tutorial. Also, we will prepare other type of evidences too.

Retraining AUGUSTUS

The main steps are:

 1. COMPILE A SET OF TRAINING AND TEST GENES
 
 2. CREATE A META PARAMETERS FILE FOR YOUR SPECIES
  
 3. MAKE AN INITIAL TRAINING
 
 4. RUN THE SCRIPT optimize_augustus.pl
 
 We will use spliced alignments of de novo assembled transcriptome short reads (RNA-Seq) and 
 spliced alignments of protein sequences against the assembled genomic sequence.
 
 You can find more detailed information in the training manual.

Also, we should prepare hints that the gene prediction tool can incorporate to the analysis. It will change the likelihood of gene structures candidates. Therefore, the algorithm will tend to predict gene structures that are in agreement with the hints. In the readme about AUGUSTUS in the RGASP assessment a detailed method is described that would produce the necessary inputs for this step. You can find more detailed information reading the AUGUSTUS tutorial.

Creating hints from RNA-Seq data

  Massive amounts of short transcriptome reads first need to be aligned to the genome.  We will assume that we have
  already aligned the reads to the genome and we have the WIG and GFF files. If you do just run the following STAR code:
  
  #build genome index
  STAR --runThreadN 4 --runMode genomeGenerate --genomeDir gindex --genomeFastaFiles genome.fasta
  
  #align reads
  STAR --runThreadN 4 --genomeDir gindex --readFilesIn rnaseq1.fq rnaseq2.fq --alignIntronMax 100000 \
  --outSAMtype BAM  SortedByCoordinate --outWigType wiggle --outWigStrand Unstranded
  
  These commands will create a file Aligned.sortedByCoord.out.bam with the alignments and a file Signal.Unique.str1.out.wig with 
  coverage per base information, among others. 
  
  1. The file coverage.wig contains a coverage graph, that contains for each base in the genome and the number of reads
     alignments that cover the position.
  
  2. The file hints.rnaseq.intron.gff contains likely intron positions, inferred from gaps in the query of the read
     alignments. Together with the intron boundaries the multiplicity (mult) is reported, which counts the number of
     alignments that support the given intron candidate, if there is more than one.
     
  Generate hints about exonic regions from the coverage graph (wig file):
  
    cat coverage.wig | wig2hints.pl --width=10 --margin=10 --minthresh=2 --minscore=4 \
    --src=W --type=ep --radius=4.5 > hints.rnaseq.ep.gff

Creating hints from cDNA

    blat -minIdentity=92 <genome.fa> <cdna.fa> <cdna.psl>

    pslCDnaFilter -maxAligns=1 cdna.psl cdna.f.psl

    blat2hints.pl --in=cdna.f.psl --out=hints.cdna.gff

Concatenate all hints from all type sources into one file:

    cat hints.cnda.gff hints.rnaseq.intron.gff hints.rnaseq.ep.gff > hints.gff

We can prepare various gene models and hints file for our genome. Later, we will use these files in Section 5.

3. Repeat modelling & masking

Repeat modelling and masking repeats is a crucial step in the workflow. We can increase the gene prediction tool speed and efficiency by masking out regions that contains non-protein coding elements. One of the most popular tools is RepeatModeler (including RepeatMasker).

The main steps are:

1. Create a Database for RepeatModeler

   <RepeatModelerPath>/BuildDatabase -name genome_of_interest genome_of_interest.fasta

2. Run RepeatModeler

   <RepeatModelerPath>/RepeatModeler -database genome_of_interest -pa 20 -LTRStruct >& run.out &
   
3. Interperting the results

   At the succesful completion of a run, two files are generated:
   
     <database_name>-families.fa  : Consensus sequences
     <database_name>-families.stk : Seed alignments
     
4. Making repeat library

   <RepeatMaskerPath>/RepeatMasker -lib <database_name>-families.fa mySequence.fa
   
5. Predicting repeats

   <RepeatMaskerPath>/RepeatMasker -small -gff --species <query species> --lib [filename for custom library] yourgenome.fasta

A few gene prediction tool can recognize masked regions if your genom sequence is upper case and the masked regios are in lower case. Therefore we highly recommend to use -small option. Later, we will need the repeats positions information in GFF fromat (use option -gff) in Section 6.

Eukaryotic genomes can be huge and repeat masking can take a lot of time (also memory). If the -pa(rallel) option is not working you can speed up this step by splitting up your genome into separate FASTA files and you can mask these files parallel. At the end you can simply combine the masked FASTA and GFF files into one.

4. Predicting tRNAs

There are not only protein coding genes in your genome of interest. If you want to see the whole picture you should predict tRNA coding genes too. tRNAscan-SE is one of the best tools for tRNA gene predictions. It is easy to run, fast and it has a webserver.

Simple usage (see more options in the manual):

 #Example 
 
  tRNAscan-SE [-options] <FASTA file(s)>

5. Predicting protein coding genes

This is the core part of the workflow. We will use an evidence driven ab initio gene prediction approach by running several gene predictor tools with various parameters. Since we want to predict thousand of protein coding genes we automatize the process but we don't have time to refine all gene models by hand. We will use more than one tool (each algorthim has it's advandage.) and in the next section we will combine these results.

We will run the following algorithms:

GeneMark-ES

New genomes can be analyzed by GeneMark-ES applying (un)supervised self-training which is an important feature of the algorithm. Also, it can take external evidences (RNA or protein) mapped to genome. So, we can use the previously prepared files (see Section 2.). We can run the gmes_petap.pl perl script to do the trainig and prediction steps in one. See the documentation for more information.

    #Example
    
     gmes_petap.pl --soft_mask --ES --evidence hints.gff --cores <number of cores> --sequence genome_of_interest.fasta

The ouput is a GTF file. However, the program can predict incomplete genes but we are not interesed in these gene models. We can filter out using the filter_genemark.R script that you can find in the repository. Next we should convert GTF into GFF using gffread.

You should run GeneMark-ES only once!

AUGUSTUS

AUGUSTUS is a program that predicts genes in eukaryotic genomic sequences. It can be run on a web server or locally. It has 2 mandatory arguments, the query file and the species. You can find more details about how to run the tool in the manual.

You can run AUGUSTUS several times using the combinations of the generic gene model, the gene model of the closest related species, retrained new species with hints that we prepared in Section 2.

Exmple runs (each case we will predict only complete genes on both strands):

  #Run 1 - using only the gene model of the closes related species
   
   augustus --strand=both --genemodel=complete --species=[related_speices] --gff3=on --codingseq=on --outfile=[out_file] genome_of_interest.fasta
   
  #Run 2 - using the gene model of the closes related species with hints
   
   augustus --strand=both --genemodel=complete --species=[related_speices] --gff3=on --codingseq=on --hintsfile=hints.gff --extrinsicCfgFile=extrinsic.ME.cfg --outfile=[out_file] genome_of_interest.fasta
   
  #Run 3 - using the generic gene model with hints
   
   augustus --strand=both --genemodel=complete --species=generic --gff3=on --codingseq=on --hintsfile=hints.E.gff --extrinsicCfgFile=extrinsic.ME.cfg --outfile=[out_file] genome_of_interest.fasta

At the end of these analyses we will use the GFF files from each run (GeneMark-ES and AUGUSTUS). You may have to uniform the GFF file format amongst the different runs.

6. Combining gene models

In this section we will use all external evidences, repeats and gene prediction results in order to merge the different gene models into one.

The EVidenceModeler (EVM) software combines gene predictions and protein and/or transcript alignments (as external evidences) into weighted consensus gene models. EVM provides a flexible and intuitive framework for combining diverse evidence types into a single automated gene structure annotation system.

Inputs to EVM include:

  • the genome sequence
  • gene predictions in GFF3 format
  • alignment data in GFF3 format
  • list of numeric weight values to be applied to each type of evidence

The weights can be configured manually!

Preparing the inputs

1. Gene predictions

We should merge all GFF files from all gene predictions into one file.

  #Example
  
   cat prediction_1.gff prediction_2.gff prediction_3.gff prediction_N.gff > gene_predictions.gff

If the separate GFF files contains headers (lines statring whit "#") you sould delete them. Other important thing is the source column (the 2nd one) of GFF files which should be unique corresponding to each run.

   #Example GFF files
   
    Scaffold2	AUGUSTUS_1	gene	7345	7746	1	-	.	ID=g1
    Scaffold2	AUGUSTUS_2	gene	7345	7746	1	-	.	ID=g1
    Scaffold2	AUGUSTUS_3	gene	7345	13968	0.07	-	.	ID=g1
    Scaffold2	GeneMark.hmm	gene	7345	7746	1	-	.	ID=g1

Here, the gene IDs are not relevant because the tool will use the coordintes to find overlaps between different models.

2. Alingments for evidences

We can prepare cDNA (assembled transcripts from RNA-seq data) alignments using GMAP/GSNAP. We can do it in two simple steps:

    #Creating genome index
    
     gmap_build -d <genome> [-k <kmer size>] <fasta_files...>
    
    #Mapping transcripts to the genome
    
     gmap -d <genome> -B 5 -t 10 -f 3 transcripts.fasta

Rename the output file as cdna.gff!

Also, we can use protein based evidences generated with Exonerate and Scipio.

     #Exonerate example
     
      exonerate --model protein2genome query_proteins.fasta target_genome.fasta
      
     #Scipio example
     
      scipio.pl [<options>] <target genome> <query protein>

See more option in the manuals (Exonerate, Scipio).

Merge then rename the output file as proteins.gff! Again, the source column (the 2nd one) of GFF files should be unique corresponding to each run.

3. Repeats

We can use the GFF output from repeat masing (see Section 3.) Rename the output as repeats.gff.

4. Weights

We can easily creat the weights file which has three columns:

  • evidence class
  • type, correspoding to the name of the tool/run that was used to generate it
  • weight

The class parameter can be one of the following:

  • ABINITIO_PREDICTION
  • PROTEIN
  • TRANSCRIPT

These are the only input types accepted by EVM. An example weights file looks like:

     TRANSCRIPT	GMAP	10
     PROTEIN	exonerate	5
     PROTEIN	Scipio	5
     ABINITIO_PREDICTION	AUGUSTUS_1	1
     ABINITIO_PREDICTION	AUGUSTUS_2	1
     ABINITIO_PREDICTION	AUGUSTUS_3	1
     ABINITIO_PREDICTION	GeneMark.hmm	1

Running EVM

Now, we have all the necessary inputs for running EVidenceModeler.

The main steps are:

  • Partitioning the Inputs - EVM will split the input dataset based on the contigs/scaffolds

    perl $EVM_HOME/EvmUtils/partition_EVM_inputs.pl --genome sotfmasked_genome.fasta \
         --gene_predictions gene_predictions.gff --protein_alignments proteins.gff \
         --transcript_alignments cdna.gff \
         --repeats repeats.gff \
         --segmentSize 100000 --overlapSize 10000 --partition_listing partitions_list.out
    
  • Generating the EVM Command Set (one command for one contig/scaffold)

    perl ${EVM_HOME}/EvmUtils/write_EVM_commands.pl --genome sotfmasked_genome.fasta --weights weights.txt \
         --gene_predictions gene_predictions.gff --protein_alignments proteins.gff \
         --transcript_alignments cdna.gff \
         --repeats repeats.gff \
         --output_file_name evm.out  --partitions partitions_list.out >  commands.list
    
    perl $EVM_HOME/EvmUtils/execute_EVM_commands.pl commands.list | tee run.log
    
  • Combining the Partitions

    perl $EVM_HOME/EvmUtils/recombine_EVM_partial_outputs.pl --partitions partitions_list.out --output_file_name evm.out
    
  • Convert the results to GFF3 Format

    perl $EVM_HOME/EvmUtils/convert_EVM_outputs_to_GFF3.pl  --partitions partitions_list.out --output evm.out  --genome sotfmasked_genome.fasta
    
  • Merge GFF files

    cat <Name of the contigs/scaffolds>/evm.out > evm.all.out
    cat <Name of the contigs/scaffolds>/evm.out.gff3 > evm.out.all.gff
    

Or you can use the runEVidenceModeler.sh script. Before running it you may have change it a little bit!

At the end of this process we can filter the results based on the number of evidences that support a gene model. For example we get rid of those models that have only 1 ab initio evidence per feature (exon/intron). You can use the filter_evm_run.R script with basic filtering options. It will allow you to filter the results based on the average weights per gene model. The inputs are the evm.out.all.gff and evm.all.out files.

Also, you can come up with your own idea for filtering the results.

Finally we will extract the CDS and protein sequences that we will use in the annotation step (Section 7.). In this workflow we don't care about the alternative splicing events. However, it can add more to the annotation.

7. Annotating protein coding genes

Now we have all the protein coding genes from Section 6. We will annotate structural and functional features of the coding sequences/proteins.

One of the most popular approach is to annotate the gene of interes based on similarity unsing BLAST. If you have to align thousands of sequnces you can use the local version and you can prepare a custom database by colleting all well annotated proteins/transcripts from closely related species or you can use the UniProtKB/Swiss-Prot database. You can align DNA to DNA or PROTEIN to DNA.

See more in the BLAST documentation how to use the command line version.

However, BALST has good balance of sensitivity and speed, it is flexible and reliable BUT the quality of the results depends on your database. It can happen there are a lot of unannotated and hypothetical elements in your databese. Also, if your databes is good you may have a lot of elemets that have poor matches etc...

Therefore, we highly recommend to scan your sequences for matches against a protein signature databases. For this purpose InterPro provides functional analysis of proteins by classifying them into families and predicting domains and important sites. You can do quick search with limited number of amino acids (40,000). If you want to analyze thousands of sequences you can download the InterProScan tool. It has a github repository but for more information on downloading, installing and running it please see the wiki page.

Finally, we can gain more information by reconstructing KEGG pathways using KAAS or BlastKOALA. Than mapping KO elements to general pathways or comparing it to the closest realted species using KEGGREST Bioconductor package or other tools.

BLAST example

Create a custom database from a multi-FASTA file of sequences:

 #Example
 
  makeblastdb –in mydb.fsa –dbtype nucl –parse_seqids	 

A BLAST search against a database requires at least a –query and –db option:

 #Example
  
  blastn –db nt –query nt.fsa –out results.out  

InterProScan example

You can run InterProScan directly from the command line (be sure you database is updated). In the following example we will use the protein sequences for the analysis. If the --applications <ANALYSES> option is not set, ALL analyses will be run.

Available analyses:

  • TIGRFAM (XX.X) : TIGRFAMs are protein families based on Hidden Markov Models or HMMs
  • SFLD (X.X) : SFLDs are protein families based on Hidden Markov Models or HMMs
  • ProDom (XXXX.X) : ProDom is a comprehensive set of protein domain families automatically generated from the UniProt Knowledge Database
  • Hamap (XXXXXX.XX) : High-quality Automated and Manual Annotation of Microbial Proteomes
  • SMART (X.X) : SMART allows the identification and analysis of domain architectures based on Hidden Markov Models or HMMs
  • CDD (X.XX) : Prediction of CDD domains in Proteins
  • ProSiteProfiles (XX.XXX) : PROSITE consists of documentation entries describing protein domains, families and functional sites as well as associated patterns and profiles to identify them
  • ProSitePatterns (XX.XXX) : PROSITE consists of documentation entries describing protein domains, families and functional sites as well as associated patterns and profiles to identify them
  • SUPERFAMILY (X.XX) : SUPERFAMILY is a database of structural and functional annotation for all proteins and genomes
  • PRINTS (XX.X) : A fingerprint is a group of conserved motifs used to characterise a protein family
  • PANTHER (X.X) : The PANTHER (Protein ANalysis THrough Evolutionary Relationships) Classification System is a unique resource that classifies genes by their functions, using published scientific experimental evidence and evolutionary relationships to - - Gene3D (X.X.X) : Structural assignment for whole genes and genomes using the CATH domain structure database
  • PIRSF (X.XX) : The PIRSF concept is being used as a guiding principle to provide comprehensive and non-overlapping clustering of UniProtKB sequences into a hierarchical order to reflect their evolutionary relationships.
  • Pfam (XX.X) : A large collection of protein families, each represented by multiple sequence alignments and hidden Markov models (HMMs)
  • Coils (X.X) : Prediction of Coiled Coil Regions in Proteins
  • MobiDBLite (X.X) : Prediction of disordered domains Regions in Proteins

Also, we switch on the GO and pathway annotation options. For further analyses or making summary of the results we can use the TSV and GFF outputs. For better interpretation use the HTML output.

  #Example
  
   interproscan.sh --output-dir <OUTPUT-DIR> --output-file-base <OUTPUT-FILE-BASE> --formats GFF,TSV,HTML --goterms --iprlookup --pathways --seqtype protein --input <INPUT-FILE-PATH>

About

Pipeline for eukaryotic genome annotation based on external evidences

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published