Bioinformatics at COMAV

Cleaning sequence reads

ngs_backbone can clean sanger, 454, illumina and solid sequences. This process usually involves vector and adaptor removal, bad quality regions trimming and short sequence filtering. There are four cleaning pipelines defined in ngs_backbone that are used depending on the platform and on the quality availability:

long reads with quality
for sanger and 454 sequences with quality information
long reads without quality
for sanger reads without quality information
for short illumina reads
for solid reads

A collection of cleaning steps are available that compose each one of these pipelines. These steps are:

adaptor removal
Each sequence is align against the adaptors found in a fasta file. The external tool used to do the matching is blast with its short sequence algorithm. If a match is found this section of the read is removed.
short adaptor removal
ngs_backbone will look for adaptors shorter than 15 bp with exact matches.
precise vector removal
If the vector and cloning site is known lucy can be used to remove the vector in a precise way.
vector removal
If only the vector is known we can use blast to remove it.
general vector removal
The reads are compared against the Univec database using blast to look for remaining vectors.
bad quality trimming
There are two algorithms used to remove the bad quality sequence extremes. If the sequence is long lucy (454 and sanger) is used for this task otherwise ngs_backbone does the job (illumina and solid).
bad quality trimming by Ns
When there are no qualities available we can infer the quality from the percent of Ns found in the sequence. This step removes regions with a high number of Ns.
low complexity masking
The regions with a low complexity can be masked by using mdust
edge removal
After all the other modules are run we can delete a fixed amount of bases from the sequence extremes
short sequence filtering
When the process for one sequence is completed a minimum length criteria is applied.

The pipelines are:

long reads with quality
adaptor removal, precise vector removal, bad quality trimming, general vector removal, low complexity masking, word removal, edge removal, and short sequence filtering
long reads without quality
general vector removal, bad quality trimming by Ns, low complexity masking, word removal, edge removal, and short sequence filtering
adaptor removal, bad quality trimming, and short sequence filtering
bad quality trimming and short sequence filtering

Input and output files

The reads to be cleaned should be in the project directory under /reads/raw/. The naming conventions should be followed by these files, especially the bit regarding to the extension and the platform. The output files will have the same names, but they will be located at /reads/cleaned/. The analysis will proceed for all sequence files found in /reads/raw, if a matching file is not found in /reads/cleaned/ a new cleaned file will be generated. If a matching file is found in /reads/cleaned/ these file will not be overwritten, so the analysis for this file will not be repeated until the file from /reads/cleaned is removed.

Configuration parameters

The configuration for the cleaning analysis is found in the Cleaning section on the ngs_backbone.conf file. The parameters are:

The blast database that will be used to look for clonning vectors.
A path to a fasta file containing the adaptors used to build the 454 library. They will be removed from the cleanend reads.
Idem for the sanger sequences
Idem for the illumina sequences
A list of words to be removed. They can be regular expressions.
Idem for the sanger sequences
Idem for the illumina sequences
edge_removal -> 454_left
A fixed number of bases to be removed from the left edge of the 454 reads.
edge_removal -> 454_right
Idem for the right edge of the 454 reads
edge_removal -> sanger_left
Idem for the left edge of the sanger reads
edge_removal -> sanger_right
Idem for the right edge of the sanger reads
edge_removal -> illumina_left
Idem for the left edge of the illumina reads
edge_removal -> illumina_right
Idem for the right edge of the illumina reads
Threshold used for the trimming of regions with a lot of Ns in sequences with no quality information available. Lower values (e.g. 1.5) are more stringent.
The minimum sequence length allowable after the cleaning is done. All sequences shorter than these values will be discarded. This is a subsection with one value for each platform 454, sanger and illumina.
lucy -> vector_settings
A path to a lucy settings file with the splice and vectors files to be used by lucy
lucy ->bracket
Look at lucy man page before changing defaults.
lucy -> window
Look at lucy man page before changing defaults.
lucy -> error
Look at lucy man page before changing defaults.

lucy settings

The lucy settings file should have the following format:

{'library1':{'vector_file':'lib1_vector.fasta', 'splice_file':'lib1_splice.fasta'},
 'library2':{'vector_file':'lib2_vector.fasta', 'splice_file':'lib2_splice.fasta'},}

In this file the paths to the vector and splice files for lucy should be stated for every library to be cleaned by lucy. The library name will be scraped from the read sequence file (that should follow the naming conventions. The vector file is just a fasta file, the information to be set in the splice file should is explained in the lucy man page.