You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
--> NOPE :( -- First result: f_measure=0.7536231884057971, recall just improved few negligible decimals, precision dropped from 91 to 80...
Investigate 550 relations in test_corpus_stats (vs 1345 ?)
Investigate the lower recall for the baseline?
Investigate further links for sentence combination
Investigate multiple dependencies (I must get the shortest one!)
[ ] Investigate UNKNOWN normalization
Investigate Shrikant's D1 features
Do feature selection for all models up to D6
Now do progressive combined sequences of all models: [D0, D1, D2, D3, D4, D5, D6]
--> when F1 drops?
(By Juanmi) Implement F-Beta scoring
Experiment with optimal Beta values and optimal number of m models
Put everything together
Other
Must Haves
Investigate possible very slight random changes performance
(By Juanmi) Do scaling independently for training & testing (?) SklSVM._preprocess(X), and definitely keep the maximums of training data for scaling on new unseen instances (which come only a few, so the proportions would be totally different)
(By Juanmi) Fix: evaluate in macro -- I'm actually evaluating on micro and the documentation in nalaf is likely wrong
Main Steps
Use combined sentences with current links
Run
D1alone with pre-selected features from D0P=61, R=16, F=25P=21, R=21, F1=21Run
D1alone with ALL features:P=54, R=0.02, F=0.04P=46, R=0.07, F1=12Do feature selection for
D1--> ~206 featuresP=94, R=14, F=25-->49 tpP=90, R=47, F=61-->38 tpMerge
D0andD1f_measure=0.7536231884057971, recall just improved few negligible decimals, precision dropped from 91 to 80...Investigate 550 relations in
test_corpus_stats(vs 1345 ?)Investigate the lower recall for the baseline?
Investigate further links for sentence combination
Investigate multiple dependencies (I must get the shortest one!)
[ ] Investigate UNKNOWN normalizationInvestigate Shrikant's D1 features
Do feature selection for all models up to
D6Now do progressive combined sequences of all models:
[D0, D1, D2, D3, D4, D5, D6](By Juanmi) Implement F-Beta scoring
Experiment with optimal
Betavalues and optimal number ofmmodelsPut everything together
Other
Must Haves
SklSVM._preprocess(X), and definitely keep the maximums of training data for scaling on new unseen instances (which come only a few, so the proportions would be totally different)Nice to Haves, in rough order of priority
[Plant, steroid, hormones, ,, brassinosteroids, (, BRs, ), ,, are, perceived, by, the, plasma, membrane, -, localized, leucine, -, rich, -, repeat, -, receptor, kinase, BRI, 1, ., Based, on, sequence, similarity, ,, we, have, identified, three, members, of, the, BRI, 1, family, ,, named, BRL, 1, ,, BRL, 2, and, BRL, 3, .]?is_enzymefrom SwisProt