git://source.jalview.org
/
jalview.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
keep it as close with archy.js as possible...
[jalview.git]
/
forester
/
java
/
src
/
org
/
forester
/
io
/
parsers
/
HmmPfamOutputParser.java
diff --git
a/forester/java/src/org/forester/io/parsers/HmmPfamOutputParser.java
b/forester/java/src/org/forester/io/parsers/HmmPfamOutputParser.java
index
eda7e55
..
96729a4
100644
(file)
--- a/
forester/java/src/org/forester/io/parsers/HmmPfamOutputParser.java
+++ b/
forester/java/src/org/forester/io/parsers/HmmPfamOutputParser.java
@@
-64,7
+64,6
@@
public final class HmmPfamOutputParser {
private final FilterType _filter_type;
private final File _input_file;
private final String _species;
private final FilterType _filter_type;
private final File _input_file;
private final String _species;
- private final String _model_type;
private double _e_value_maximum;
private Map<String, String> _individual_domain_score_cutoffs;
private boolean _ignore_dufs;
private double _e_value_maximum;
private Map<String, String> _individual_domain_score_cutoffs;
private boolean _ignore_dufs;
@@
-93,7
+92,6
@@
public final class HmmPfamOutputParser {
public HmmPfamOutputParser( final File input_file, final String species, final String model_type ) {
_input_file = input_file;
_species = species;
public HmmPfamOutputParser( final File input_file, final String species, final String model_type ) {
_input_file = input_file;
_species = species;
- _model_type = model_type;
_filter = null;
_filter_type = FilterType.NONE;
init();
_filter = null;
_filter_type = FilterType.NONE;
init();
@@
-101,12
+99,10
@@
public final class HmmPfamOutputParser {
public HmmPfamOutputParser( final File input_file,
final String species,
public HmmPfamOutputParser( final File input_file,
final String species,
- final String model_type,
final Set<String> filter,
final FilterType filter_type ) {
_input_file = input_file;
_species = species;
final Set<String> filter,
final FilterType filter_type ) {
_input_file = input_file;
_species = species;
- _model_type = model_type;
_filter = filter;
_filter_type = filter_type;
init();
_filter = filter;
_filter_type = filter_type;
init();
@@
-218,10
+214,6
@@
public final class HmmPfamOutputParser {
return _max_allowed_overlap;
}
return _max_allowed_overlap;
}
- private String getModelType() {
- return _model_type;
- }
-
public int getProteinsEncountered() {
return _proteins_encountered;
}
public int getProteinsEncountered() {
return _proteins_encountered;
}
@@
-341,8
+333,8
@@
public final class HmmPfamOutputParser {
}
else if ( isVerbose() ) {
ForesterUtil.printWarningMessage( getClass().getName(), "query \"" + query
}
else if ( isVerbose() ) {
ForesterUtil.printWarningMessage( getClass().getName(), "query \"" + query
- + "\" is not unique [line " + line_number + "] in ["
- + getInputFile().getCanonicalPath() + "]" );
+ + "\" is not unique [line " + line_number + "] in ["
+ + getInputFile().getCanonicalPath() + "]" );
}
}
else {
}
}
else {
@@
-436,38
+428,36
@@
public final class HmmPfamOutputParser {
int to = -1;
double e_value = -1;
double score = -1;
int to = -1;
double e_value = -1;
double score = -1;
- boolean is_complete_hmm_match = false;
- boolean is_complete_query_match = false;
try {
from = Integer.valueOf( from_str ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse seq-f from \"" + line + "\" [line " + line_number
try {
from = Integer.valueOf( from_str ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse seq-f from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
try {
to = Integer.valueOf( to_str ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse seq-t from \"" + line + "\" [line " + line_number
}
try {
to = Integer.valueOf( to_str ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse seq-t from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
try {
score = Double.valueOf( score_str ).doubleValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse score from \"" + line + "\" [line " + line_number
}
try {
score = Double.valueOf( score_str ).doubleValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse score from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
try {
e_value = Double.valueOf( e_value_str ).doubleValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse E-value from \"" + line + "\" [line " + line_number
}
try {
e_value = Double.valueOf( e_value_str ).doubleValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse E-value from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
if ( hmm_match_str.equals( "[]" ) ) {
}
if ( hmm_match_str.equals( "[]" ) ) {
- is_complete_hmm_match = true;
+ //is_complete_hmm_match = true;
}
else if ( !( hmm_match_str.equals( ".]" ) || hmm_match_str.equals( "[." ) || hmm_match_str
.equals( ".." ) ) ) {
}
else if ( !( hmm_match_str.equals( ".]" ) || hmm_match_str.equals( "[." ) || hmm_match_str
.equals( ".." ) ) ) {
@@
-475,7
+465,7
@@
public final class HmmPfamOutputParser {
+ line_number + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
if ( query_match_str.equals( ".." ) ) {
+ line_number + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
if ( query_match_str.equals( ".." ) ) {
- is_complete_query_match = true;
+ // is_complete_query_match = true;
}
else if ( !( query_match_str.equals( ".]" ) || query_match_str.equals( "[." ) || query_match_str
.equals( "[]" ) ) ) {
}
else if ( !( query_match_str.equals( ".]" ) || query_match_str.equals( "[." ) || query_match_str
.equals( "[]" ) ) ) {
@@
-496,14
+486,14
@@
public final class HmmPfamOutputParser {
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse domain number from \"" + line + "\" [line " + line_number
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse domain number from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
try {
total = Integer.valueOf( ( total_str ) ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse domain count from \"" + line + "\" [line " + line_number
}
try {
total = Integer.valueOf( ( total_str ) ).intValue();
}
catch ( final NumberFormatException e ) {
throw new IOException( "could not parse domain count from \"" + line + "\" [line " + line_number
- + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
++_domains_encountered;
boolean failed_cutoff = false;
}
++_domains_encountered;
boolean failed_cutoff = false;
@@
-516,7
+506,7
@@
public final class HmmPfamOutputParser {
}
else {
throw new IOException( "could not find a score cutoff value for domain id \"" + id
}
else {
throw new IOException( "could not find a score cutoff value for domain id \"" + id
- + "\" [line " + line_number + "] in [" + getInputFile().getCanonicalPath() + "]" );
+ + "\" [line " + line_number + "] in [" + getInputFile().getCanonicalPath() + "]" );
}
}
final String uc_id = id.toUpperCase();
}
}
final String uc_id = id.toUpperCase();
@@
-623,8
+613,8
@@
public final class HmmPfamOutputParser {
/**
* To ignore domains which are completely engulfed by domains (individual
* ones or stretches of overlapping ones) with better support values.
/**
* To ignore domains which are completely engulfed by domains (individual
* ones or stretches of overlapping ones) with better support values.
- *
- *
+ *
+ *
* @param ignored_engulfed_domains
*/
public void setIgnoreEngulfedDomains( final boolean ignore_engulfed_domains ) {
* @param ignored_engulfed_domains
*/
public void setIgnoreEngulfedDomains( final boolean ignore_engulfed_domains ) {
@@
-638,7
+628,7
@@
public final class HmmPfamOutputParser {
/**
* Sets the individual domain score cutoff values (for example, gathering
* thresholds from Pfam). Domain ids are the keys, cutoffs the values.
/**
* Sets the individual domain score cutoff values (for example, gathering
* thresholds from Pfam). Domain ids are the keys, cutoffs the values.
- *
+ *
* @param individual_domain_score_cutoffs
*/
public void setIndividualDomainScoreCutoffs( final Map<String, String> individual_domain_score_cutoffs ) {
* @param individual_domain_score_cutoffs
*/
public void setIndividualDomainScoreCutoffs( final Map<String, String> individual_domain_score_cutoffs ) {