## Tophat Mapping Report from the logs TophatMappingReport(){ echo ' devtools::source_url("https://dl.dropboxusercontent.com/u/113630701/datautils/R/core_commons.R") devtools::source_url("https://dl.dropboxusercontent.com/u/113630701/datautils/R/ggplot_commons.R") parseAlgnSummary_T2_0_11 <- function(alignSummary){ #alignSummary="/projects/bioinfo/holger/projects/marta_rnaseq/human_leipzig/mapping/S5382_aRG_1b_rep1/align_summary.txt" algnData <- readLines(alignSummary) data.frame( condition=basename(dirname(alignSummary)), num_reads=as.numeric(str_match(algnData[2], " ([0-9]*$)")[,2]), mapped_reads=as.numeric(str_match(algnData[3], ":[ ]*([0-9]*) ")[,2][1]) ) %>% transform(mapping_efficiency=100*mapped_reads/num_reads) } algnSummary <- ldply(list.files(".", "align_summary.txt", full.names=TRUE, recursive=T), parseAlgnSummary_T2_0_11, .progress="text") write.delim(algnSummary, file="tophat_mapping_stats.txt") scale_fill_discrete <- function (...){ scale_color_brewer(..., type = "seq", palette="Set1", "fill", na.value = "grey50") } projectName=basename(dirname(getwd())) devtools::source_url("https://raw.githubusercontent.com/holgerbrandl/mdreport/master/R/mdreport-package.r") md_new(paste("Mapping Summary for ", projectName)) md_plot(ggplot(algnSummary, aes(condition, mapping_efficiency)) + geom_bar(stat="identity") +coord_flip() + ylim(0,100) + ggtitle("mapping efficiency")) md_plot(ggplot(algnSummary, aes(condition, num_reads)) + geom_bar(stat="identity") + coord_flip() + ggtitle("read counts") +scale_y_continuous(labels=comma)) md_plot(ggplot(algnSummary, aes(condition, mapped_reads)) + geom_bar(stat="identity") + coord_flip() + ggtitle("alignments counts") +scale_y_continuous(labels=comma)) #ggplot(melt(algnSummary), aes(condition, value)) + geom_bar(stat="identity") +facet_wrap(~variable, scales="free") + ggtitle("mapping summary") + scale_y_continuous(labels=comma) + theme(axis.text.x=element_text(angle=90, hjust=0)) #ggsave2(w=10, h=10, p="mapstats") md_report("tophat_mapping_report", open=F) ' | R -q --vanilla } export -f TophatMappingReport #### Bowtie Mapping Report from the logs Bowtie2MappingReport(){ echo ' devtools::source_url("http://dl.dropbox.com/u/113630701/rlibs/base-commons.R") logSuffix=".logs" parseAlgnSummary <- function(alignSummary){ #alignSummary="./H2Az_Rep1_Lane1_Lib4454.bowtie.log" algnData <- readLines(alignSummary) data.frame( condition=trimEnd(basename(alignSummary), logSuffix), num_reads=as.numeric(str_split_fixed(algnData[3], " ", 2)[1]), unique_mappers=as.numeric(str_split_fixed(str_trim(algnData[6]), " ", 2)[1]), mapping_efficiency=as.numeric(str_replace(str_split_fixed(algnData[8], " ", 2)[1], "%", "")), multi_mappers=as.numeric(str_split_fixed(str_trim(algnData[7]), " ", 2)[1]) ) } mapStats <- ldply(list.files(".", logSuffix, full.names=TRUE, recursive=T), parseAlgnSummary, .progress="text") write.delim(mapStats, file="mapStats.txt") ggplot(melt(mapStats), aes(condition, value)) + geom_bar(stat="identity") +facet_wrap(~variable, scales="free") + ggtitle("mapping summary") + scale_y_continuous(labels=comma) + theme(axis.text.x=element_text(angle=90, hjust=0)) ggsave2(w=10, h=10, p="mapstats") ggplot(mapStats, aes(condition, mapping_efficiency)) + geom_bar(stat="identity") +coord_flip() + ylim(0,100) + ggtitle("mapping efficiency") ggsave2(p="mapstats") ggplot(mapStats, aes(condition, num_reads)) + geom_bar(stat="identity") + coord_flip() + ggtitle("read counts") ggsave2(p="mapstats") ggplot(mapStats, aes(condition, unique_mappers)) + geom_bar(stat="identity") + coord_flip() + ggtitle("unique alignment") + scale_fill_discrete() ggsave2(p="mapstats") ' | R --vanilla } export -f Bowtie2MappingReport ### Create a cuffdb on a network of lustre file-systen MakeCuffDB() { if [ $# -ne 2 ]; then echo "Usage: MakeCuffDB <gtffile> <genomebuild>"; return; fi echo ' devtools::source_url("http://dl.dropbox.com/u/113630701/rlibs/base-commons.R") options(width=150) require.auto(cummeRbund) createCuffDbTrickyDisk <- function(dbDir, gtfFile, genome, ...){ tmpdir <- tempfile() system(paste("cp -r", dbDir, tmpdir)) oldWD <- getwd() setwd(tmpdir) cuff <- readCufflinks(rebuild=T, gtf=gtfFile, genome="mm10", ...) # cuff <- readCufflinks(gtf=gtfFile, genome="mm10", rebuild=T) system(paste("cp cuffData.db", dbDir)) system(paste("rm -r", tmpdir)) setwd(oldWD) return(cuff) } gtfFile=commandArgs(TRUE)[1] genomeBuild=commandArgs(TRUE)[2] createCuffDbTrickyDisk(getwd(), gtfFile, genomeBuild) ' | R -q --no-save --no-restore --args $1 $2 } export -f MakeCuffDB CountSeqs(){ for fasta in $*; do grep ">" $fasta | wc -l | sed 's/^/'$fasta':\t/g' done } export -f CountSeqs CountFastqGzReads() { zcat $1 | grep "^+$" | wc -l | sed -e s/^/num_reads,/; }; export -f CountFastqGzReads IndexBams(){ for bamFile in $1; do sem -j5 -no-notice samtools index $bamFile; done sem -no-notice --wait } export -f IndexBams ##http://www.biostars.org/p/16471/ ## estimate blast progress for fasta-query files. Result files are assumed to have the fasta id in column 1 BlastProgress(){ if [ $# -eq 0 ]; then echo "Usage: BlastProgress <blast_query_fasta>+"; return; fi for query in $* ; do if [ ! -f $query ]; then >&2 echo "Error: $query is not a file" return; fi blast=$query.blast.out if [ ! -f $blast ]; then >&2 echo "Warning: Could not find blast output file '$blast'" # return; fi # echo "the blast out is: "$blast #echo "the fasta query is: "$query #curquery=$(tail -1 $blast | cut -f 1) # http://tldp.org/LDP/abs/html/fto.html if [ -s $blast ]; then # echo "file exists and has non-zero size" curquery=$(tail -1 $blast | cut -f 1) curline=$(grep -n $curquery"$" $query | cut -f 1 -d ':') else # echo "file does not yet exist or is empty" curline=0 fi nblines=$(wc -l $query | cut -f 1 -d " ") percent=$(echo "($curline/$nblines) *100" | bc -l | cut -c 1-4) echo "Approximately $percent % of $query were processed." done } export -f BlastProgress