X-Git-Url: http://source.jalview.org/gitweb/?a=blobdiff_plain;f=forester%2Fjava%2Fsrc%2Forg%2Fforester%2Fapplication%2Fsurfacing.java;h=62c6981f81c2187c15dcb71df1e5dd1def92700e;hb=ffc6085cfd61eabca43b39c0423881b6f4fa7706;hp=b795c3e5cd1327675c350419a6dded8dc8b86158;hpb=9c8c68df524fe97f22fd19bec72dba51b06fcf43;p=jalview.git diff --git a/forester/java/src/org/forester/application/surfacing.java b/forester/java/src/org/forester/application/surfacing.java index b795c3e..62c6981 100644 --- a/forester/java/src/org/forester/application/surfacing.java +++ b/forester/java/src/org/forester/application/surfacing.java @@ -33,9 +33,11 @@ import java.io.IOException; import java.io.Writer; import java.util.ArrayList; import java.util.Date; +import java.util.HashMap; import java.util.HashSet; import java.util.List; import java.util.Map; +import java.util.Map.Entry; import java.util.Set; import java.util.SortedMap; import java.util.SortedSet; @@ -97,6 +99,7 @@ import org.forester.util.ForesterUtil; public class surfacing { + private static final int MINIMAL_NUMBER_OF_SIMILARITIES_FOR_SPLITTING = 1000; public final static String DOMAIN_COMBINITONS_OUTPUT_OPTION_FOR_GRAPH_ANALYSIS = "graph_analysis_out"; public final static String DOMAIN_COMBINITONS_OUTPUTFILE_SUFFIX_FOR_GRAPH_ANALYSIS = "_dc.dot"; public final static String PARSIMONY_OUTPUT_FITCH_PRESENT_BC_OUTPUTFILE_SUFFIX_FOR_GRAPH_ANALYSIS = "_fitch_present_dc.dot"; @@ -231,8 +234,8 @@ public class surfacing { final static private String INPUT_SPECIES_TREE_OPTION = "species_tree"; final static private String SEQ_EXTRACT_OPTION = "prot_extract"; final static private char SEPARATOR_FOR_INPUT_VALUES = '#'; - final static private String PRG_VERSION = "2.200"; - final static private String PRG_DATE = "2011.10.18"; + final static private String PRG_VERSION = "2.210"; + final static private String PRG_DATE = "2012.02.21"; final static private String E_MAIL = "czmasek@burnham.org"; final static private String WWW = "www.phylosoft.org/forester/applications/surfacing"; final static private boolean IGNORE_DUFS_DEFAULT = true; @@ -275,6 +278,13 @@ public class surfacing { public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_OUTPUT_SUFFIX = "_indep_dc_gains_fitch_lists.txt"; public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_FOR_GO_MAPPING_OUTPUT_SUFFIX = "_indep_dc_gains_fitch_lists_for_go_mapping.txt"; public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_FOR_GO_MAPPING_OUTPUT_UNIQUE_SUFFIX = "_indep_dc_gains_fitch_lists_for_go_mapping_unique.txt"; + public static final String LIMIT_SPEC_FOR_PROT_EX = null; // e.g. "HUMAN"; set to null for not using this feature (default). + public static final String BINARY_DOMAIN_COMBINATIONS_PARSIMONY_TREE_OUTPUT_SUFFIX_FITCH_MAPPED = "_dc_MAPPED_secondary_features_fitch" + + ForesterConstants.PHYLO_XML_SUFFIX; + public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_COUNTS_MAPPED_OUTPUT_SUFFIX = "_indep_dc_gains_fitch_counts_MAPPED.txt"; + public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_MAPPED_OUTPUT_SUFFIX = "_indep_dc_gains_fitch_lists_MAPPED.txt"; + public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_FOR_GO_MAPPING_MAPPED_OUTPUT_SUFFIX = "_indep_dc_gains_fitch_lists_for_go_mapping_MAPPED.txt"; + public static final String INDEPENDENT_DC_GAINS_FITCH_PARS_DC_FOR_GO_MAPPING_MAPPED_OUTPUT_UNIQUE_SUFFIX = "_indep_dc_gains_fitch_lists_for_go_mapping_unique_MAPPED.txt"; private static void checkWriteabilityForPairwiseComparisons( final PrintableDomainSimilarity.PRINT_OPTION domain_similarity_print_option, final String[][] input_file_properties, @@ -1721,6 +1731,7 @@ public class surfacing { catch ( final IOException e2 ) { ForesterUtil.fatalError( surfacing.PRG_NAME, e2.getMessage() ); } + final DescriptiveStatistics protein_coverage_stats = new BasicDescriptiveStatistics(); final DescriptiveStatistics all_genomes_domains_per_potein_stats = new BasicDescriptiveStatistics(); final SortedMap all_genomes_domains_per_potein_histo = new TreeMap(); final SortedSet domains_which_are_always_single = new TreeSet(); @@ -1748,6 +1759,7 @@ public class surfacing { catch ( final IOException e3 ) { e3.printStackTrace(); } + // Main loop: for( int i = 0; i < number_of_genomes; ++i ) { System.out.println(); System.out.println( ( i + 1 ) + "/" + number_of_genomes ); @@ -1775,12 +1787,14 @@ public class surfacing { input_file_properties[ i ][ 1 ], filter, filter_type, - ind_score_cutoff ); + ind_score_cutoff, + true ); } else { parser = new HmmscanPerDomainTableParser( new File( input_file_properties[ i ][ 0 ] ), input_file_properties[ i ][ 1 ], - ind_score_cutoff ); + ind_score_cutoff, + true ); } if ( e_value_max >= 0.0 ) { parser.setEValueMaximum( e_value_max ); @@ -1811,10 +1825,16 @@ public class surfacing { System.out.println( "Domains ignored due to virus like id: " ); ForesterUtil.printCountingMap( parser.getDomainsIgnoredDueToVirusLikeIdCountsMap() ); } + final double coverage = ( double ) protein_list.size() / parser.getProteinsEncountered(); + protein_coverage_stats.addValue( coverage ); System.out.println( "Number of proteins encountered : " + parser.getProteinsEncountered() ); log( "Number of proteins encountered : " + parser.getProteinsEncountered(), log_writer ); System.out.println( "Number of proteins stored : " + protein_list.size() ); log( "Number of proteins stored : " + protein_list.size(), log_writer ); + System.out.println( "Coverage : " + + ForesterUtil.roundToInt( 100.0 * coverage ) + "%" ); + log( "Coverage : " + ForesterUtil.roundToInt( 100.0 * coverage ) + + "%", log_writer ); System.out.println( "Domains encountered : " + parser.getDomainsEncountered() ); log( "Domains encountered : " + parser.getDomainsEncountered(), log_writer ); System.out.println( "Domains stored : " + parser.getDomainsStored() ); @@ -1894,7 +1914,7 @@ public class surfacing { catch ( final IOException e ) { ForesterUtil.fatalError( surfacing.PRG_NAME, e.toString() ); } - SurfacingUtil.domainsPerProteinsStatistics( input_file_properties[ i ][ 0 ], + SurfacingUtil.domainsPerProteinsStatistics( input_file_properties[ i ][ 1 ], protein_list, all_genomes_domains_per_potein_stats, all_genomes_domains_per_potein_histo, @@ -1933,7 +1953,8 @@ public class surfacing { SurfacingUtil.extractProteinNames( protein_list, query_domain_ids_array[ j ], query_domains_writer_ary[ j ], - "\t" ); + "\t", + LIMIT_SPEC_FOR_PROT_EX ); query_domains_writer_ary[ j ].flush(); } catch ( final IOException e ) { @@ -1952,17 +1973,6 @@ public class surfacing { } System.gc(); } // for( int i = 0; i < number_of_genomes; ++i ) { - try { - per_genome_domain_promiscuity_statistics_writer.flush(); - per_genome_domain_promiscuity_statistics_writer.close(); - dc_data_writer.flush(); - dc_data_writer.close(); - log_writer.flush(); - log_writer.close(); - } - catch ( final IOException e2 ) { - ForesterUtil.fatalError( surfacing.PRG_NAME, e2.getLocalizedMessage() ); - } ForesterUtil.programMessage( PRG_NAME, "Wrote domain promiscuities to: " + per_genome_domain_promiscuity_statistics_file ); // @@ -1981,8 +1991,8 @@ public class surfacing { domains_per_potein_stats_writer.write( "\t" ); domains_per_potein_stats_writer.write( all_genomes_domains_per_potein_stats.getMax() + "" ); domains_per_potein_stats_writer.write( "\n" ); - domains_per_potein_stats_writer.flush(); domains_per_potein_stats_writer.close(); + printOutPercentageOfMultidomainProteins( all_genomes_domains_per_potein_histo, log_writer ); ForesterUtil.map2file( new File( out_dir + ForesterUtil.FILE_SEPARATOR + output_file + "__all_genomes_domains_per_potein_histo.txt" ), all_genomes_domains_per_potein_histo, "\t", "\n" ); ForesterUtil.collection2file( new File( out_dir + ForesterUtil.FILE_SEPARATOR + output_file @@ -1991,6 +2001,16 @@ public class surfacing { + "__domains_single_or_combined.txt" ), domains_which_are_sometimes_single_sometimes_not, "\n" ); ForesterUtil.collection2file( new File( out_dir + ForesterUtil.FILE_SEPARATOR + output_file + "__domains_always_combined.txt" ), domains_which_never_single, "\n" ); + ForesterUtil.programMessage( PRG_NAME, + "Average of proteins with a least one domain assigned: " + + ( 100 * protein_coverage_stats.arithmeticMean() ) + "% (+/-" + + ( 100 * protein_coverage_stats.sampleStandardDeviation() ) + "%)" ); + ForesterUtil.programMessage( PRG_NAME, "Range of proteins with a least one domain assigned: " + 100 + * protein_coverage_stats.getMin() + "%-" + 100 * protein_coverage_stats.getMax() + "%" ); + log( "Average of prot with a least one dom assigned : " + ( 100 * protein_coverage_stats.arithmeticMean() ) + + "% (+/-" + ( 100 * protein_coverage_stats.sampleStandardDeviation() ) + "%)", log_writer ); + log( "Range of prot with a least one dom assigned : " + 100 * protein_coverage_stats.getMin() + "%-" + + 100 * protein_coverage_stats.getMax() + "%", log_writer ); } catch ( final IOException e2 ) { ForesterUtil.fatalError( surfacing.PRG_NAME, e2.getLocalizedMessage() ); @@ -2005,6 +2025,14 @@ public class surfacing { } } } + try { + per_genome_domain_promiscuity_statistics_writer.close(); + dc_data_writer.close(); + log_writer.close(); + } + catch ( final IOException e2 ) { + ForesterUtil.fatalError( surfacing.PRG_NAME, e2.getLocalizedMessage() ); + } if ( PERFORM_DOMAIN_LENGTH_ANALYSIS ) { try { SurfacingUtil.executeDomainLengthAnalysis( input_file_properties, @@ -2055,11 +2083,19 @@ public class surfacing { DescriptiveStatistics pw_stats = null; try { String my_outfile = output_file.toString(); - if ( !my_outfile.endsWith( ".html" ) ) { + Map split_writers = null; + Writer writer = null; + if ( similarities.size() > MINIMAL_NUMBER_OF_SIMILARITIES_FOR_SPLITTING ) { + if ( my_outfile.endsWith( ".html" ) ) { + my_outfile = my_outfile.substring( 0, my_outfile.length() - 5 ); + } + split_writers = new HashMap(); + createSplitWriters( out_dir, my_outfile, split_writers ); + } + else if ( !my_outfile.endsWith( ".html" ) ) { my_outfile += ".html"; + writer = new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile ) ); } - final Writer writer = new BufferedWriter( new FileWriter( out_dir == null ? my_outfile : out_dir - + ForesterUtil.FILE_SEPARATOR + my_outfile ) ); List species_order = null; if ( species_matrix ) { species_order = new ArrayList(); @@ -2079,6 +2115,7 @@ public class surfacing { .writeDomainSimilaritiesToFile( html_desc, new StringBuilder( number_of_genomes + " genomes" ), writer, + split_writers, similarities, number_of_genomes == 2, species_order, @@ -2345,6 +2382,76 @@ public class surfacing { System.out.println(); } + private static void createSplitWriters( final File out_dir, + final String my_outfile, + final Map split_writers ) throws IOException { + split_writers.put( 'a', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_A.html" ) ) ); + split_writers.put( 'b', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_B.html" ) ) ); + split_writers.put( 'c', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_C.html" ) ) ); + split_writers.put( 'd', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_D.html" ) ) ); + split_writers.put( 'e', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_E.html" ) ) ); + split_writers.put( 'f', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_F.html" ) ) ); + split_writers.put( 'g', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_G.html" ) ) ); + split_writers.put( 'h', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_H.html" ) ) ); + split_writers.put( 'i', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_I.html" ) ) ); + split_writers.put( 'j', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_J.html" ) ) ); + split_writers.put( 'k', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_K.html" ) ) ); + split_writers.put( 'l', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_L.html" ) ) ); + split_writers.put( 'm', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_M.html" ) ) ); + split_writers.put( 'n', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_N.html" ) ) ); + split_writers.put( 'o', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_O.html" ) ) ); + split_writers.put( 'p', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_P.html" ) ) ); + split_writers.put( 'q', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_Q.html" ) ) ); + split_writers.put( 'r', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_R.html" ) ) ); + split_writers.put( 's', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_S.html" ) ) ); + split_writers.put( 't', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_T.html" ) ) ); + split_writers.put( 'u', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_U.html" ) ) ); + split_writers.put( 'v', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_V.html" ) ) ); + split_writers.put( 'w', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_W.html" ) ) ); + split_writers.put( 'x', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_X.html" ) ) ); + split_writers.put( 'y', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_Y.html" ) ) ); + split_writers.put( 'z', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_Z.html" ) ) ); + split_writers.put( '0', new BufferedWriter( new FileWriter( out_dir + ForesterUtil.FILE_SEPARATOR + my_outfile + + "_domains_0.html" ) ) ); + } + + private static void printOutPercentageOfMultidomainProteins( final SortedMap all_genomes_domains_per_potein_histo, + final Writer log_writer ) { + int sum = 0; + for( final Entry entry : all_genomes_domains_per_potein_histo.entrySet() ) { + sum += entry.getValue(); + } + final double percentage = 100.0 * ( sum - all_genomes_domains_per_potein_histo.get( 1 ) ) / sum; + ForesterUtil.programMessage( PRG_NAME, "Percentage of multidomain proteins: " + percentage + "%" ); + log( "Percentage of multidomain proteins: : " + percentage + "%", log_writer ); + } + private static void preparePhylogenyForParsimonyAnalyses( final Phylogeny intree, final String[][] input_file_properties ) { final String[] genomes = new String[ input_file_properties.length ]; @@ -2361,12 +2468,21 @@ public class surfacing { final PhylogenyNode n = it.next(); if ( ForesterUtil.isEmpty( n.getName() ) ) { if ( n.getNodeData().isHasTaxonomy() + && !ForesterUtil.isEmpty( n.getNodeData().getTaxonomy().getTaxonomyCode() ) ) { + n.setName( n.getNodeData().getTaxonomy().getTaxonomyCode() ); + } + else if ( n.getNodeData().isHasTaxonomy() && !ForesterUtil.isEmpty( n.getNodeData().getTaxonomy().getScientificName() ) ) { n.setName( n.getNodeData().getTaxonomy().getScientificName() ); } + else if ( n.getNodeData().isHasTaxonomy() + && !ForesterUtil.isEmpty( n.getNodeData().getTaxonomy().getCommonName() ) ) { + n.setName( n.getNodeData().getTaxonomy().getCommonName() ); + } else { - ForesterUtil.fatalError( surfacing.PRG_NAME, - "node without both name and scientific taxonomy name found" ); + ForesterUtil + .fatalError( surfacing.PRG_NAME, + "node with no name, scientific name, common name, or taxonomy code present" ); } } } @@ -2680,7 +2796,11 @@ public class surfacing { SurfacingUtil.checkForOutputFileWriteability( out ); try { final Writer proteins_file_writer = new BufferedWriter( new FileWriter( out ) ); - SurfacingUtil.extractProteinNames( protein_lists_per_species, domain, proteins_file_writer, "\t" ); + SurfacingUtil.extractProteinNames( protein_lists_per_species, + domain, + proteins_file_writer, + "\t", + LIMIT_SPEC_FOR_PROT_EX ); proteins_file_writer.close(); } catch ( final IOException e ) {