From 651ba59f95348fe52f53e6ddeb8000df37cfdbf3 Mon Sep 17 00:00:00 2001
From: bbauvin <baptiste.bauvin@centrale-marseille.fr>
Date: Mon, 25 Sep 2017 10:20:14 -0400
Subject: [PATCH] Re-using random search for late weighted linear, added Readme

---
 Code/MonoMutliViewClassifiers/ExecClassif.py  |   17 -
 .../Monoview/ExecClassifMonoView.py           |    3 -
 .../MonoviewClassifiers/Adaboost.py           |    2 +-
 .../EarlyFusionPackage/WeightedLinear.py      |    2 +-
 Code/MonoMutliViewClassifiers/Readme.md       |   88 ++
 ...-cq-hist_lss-hist_phog-hist-awaexp-LOG.log | 1266 +++++++++++++++++
 6 files changed, 1356 insertions(+), 22 deletions(-)
 create mode 100644 Code/MonoMutliViewClassifiers/Readme.md

diff --git a/Code/MonoMutliViewClassifiers/ExecClassif.py b/Code/MonoMutliViewClassifiers/ExecClassif.py
index f16a1e46..463439d6 100644
--- a/Code/MonoMutliViewClassifiers/ExecClassif.py
+++ b/Code/MonoMutliViewClassifiers/ExecClassif.py
@@ -330,23 +330,6 @@ def initMultiviewArguments(args, benchmark, views, viewsIndices, accuracies, cla
     return argumentDictionaries
 
 
-# def analyzeLabels(labelsArrays, realLabels, classifiersNames):
-#     nbClassifiers = len(classifiersNames)
-#     nbExamples = realLabels.shape[0]
-#     nbIter = nbExamples/nbClassifiers
-#     data = np.zeros((nbExamples, nbClassifiers*nbIter))
-#     tempData = np.array([labelsArray == realLabels for labelsArray in np.transpose(labelsArrays)]).astype(int)
-#     for classifierIndex in range(nbClassifiers):
-#         for iterIndex in range(nbIter):
-#             data[:,classifierIndex*nbIter+iterIndex] = tempData[classifierIndex,:]
-#     fig, ax = plt.subplots()
-#     cax = ax.imshow(data, interpolation='nearest', cmap=cm.coolwarm)
-#     ax.set_title('Error on examples depending on the classifier')
-#     cbar = fig.colorbar(cax, ticks=[0, 1])
-#     cbar.ax.set_yticklabels(['Wrong', ' Right'])
-#     fig.savefig("Results/"+time.strftime("%Y%m%d-%H%M%S")+"error_analysis.png")
-
-
 parser = argparse.ArgumentParser(
     description='This file is used to benchmark the accuracies fo multiple classification algorithm on multiview data.',
     formatter_class=argparse.ArgumentDefaultsHelpFormatter)
diff --git a/Code/MonoMutliViewClassifiers/Monoview/ExecClassifMonoView.py b/Code/MonoMutliViewClassifiers/Monoview/ExecClassifMonoView.py
index 327974a8..214b34c3 100644
--- a/Code/MonoMutliViewClassifiers/Monoview/ExecClassifMonoView.py
+++ b/Code/MonoMutliViewClassifiers/Monoview/ExecClassifMonoView.py
@@ -68,11 +68,8 @@ def ExecMonoview(X, Y, name, labelsNames, learningRate, nbFolds, nbCores, databa
         # Calculate Train/Test data
         logging.debug("Start:\t Determine Train/Test split"+" for iteration "+str(iterationStat+1))
         testIndices = ClassifMonoView.splitDataset(Y, nbClass, learningRate, datasetLength)
-        print "fromage"
         trainIndices = [i for i in range(datasetLength) if i not in testIndices]
-        print "jqmbon"
         X_train = extractSubset(X,trainIndices)
-        print "poulet"
         X_test = extractSubset(X,testIndices)
         y_train = Y[trainIndices]
         y_test = Y[testIndices]
diff --git a/Code/MonoMutliViewClassifiers/MonoviewClassifiers/Adaboost.py b/Code/MonoMutliViewClassifiers/MonoviewClassifiers/Adaboost.py
index e1d2e0de..4e9b1c76 100644
--- a/Code/MonoMutliViewClassifiers/MonoviewClassifiers/Adaboost.py
+++ b/Code/MonoMutliViewClassifiers/MonoviewClassifiers/Adaboost.py
@@ -33,7 +33,7 @@ def gridSearch(X_train, y_train, nbFolds=4, metric=["accuracy_score", None], nIt
     else:
         metricKWARGS = {}
     scorer = metricModule.get_scorer(**metricKWARGS)
-    grid = RandomizedSearchCV(pipeline, n_iter=nIter, param_distributions=param,refit=True,n_jobs=nbCores,scoring=scorer,cv=nbFolds)
+    grid = RandomizedSearchCV(pipeline, n_iter=nIter, param_distributions=param, refit=True, n_jobs=nbCores, scoring=scorer, cv=nbFolds)
     detector = grid.fit(X_train, y_train)
     desc_estimators = [detector.best_params_["classifier__n_estimators"],
                        detector.best_params_["classifier__base_estimator"]]
diff --git a/Code/MonoMutliViewClassifiers/Multiview/Fusion/Methods/EarlyFusionPackage/WeightedLinear.py b/Code/MonoMutliViewClassifiers/Multiview/Fusion/Methods/EarlyFusionPackage/WeightedLinear.py
index 0f49fc3b..97c0ac4a 100644
--- a/Code/MonoMutliViewClassifiers/Multiview/Fusion/Methods/EarlyFusionPackage/WeightedLinear.py
+++ b/Code/MonoMutliViewClassifiers/Multiview/Fusion/Methods/EarlyFusionPackage/WeightedLinear.py
@@ -32,7 +32,7 @@ def gridSearch(DATASET, classificationKWARGS, trainIndices, nIter=30, viewsIndic
             if accuracy > bestScore:
                 bestScore = accuracy
                 bestConfig = normalizedArray
-        return [np.array([1.0 for i in range(nbView)])]
+        return [bestConfig]
 
 
 class WeightedLinear(EarlyFusionClassifier):
diff --git a/Code/MonoMutliViewClassifiers/Readme.md b/Code/MonoMutliViewClassifiers/Readme.md
new file mode 100644
index 00000000..cbe3fcd9
--- /dev/null
+++ b/Code/MonoMutliViewClassifiers/Readme.md
@@ -0,0 +1,88 @@
+# Project Title
+
+One Paragraph of project description goes here
+
+## Getting Started
+
+These instructions will get you a copy of the project up and running on your local machine for development and testing purposes. See deployment for notes on how to deploy the project on a live system.
+
+### Prerequisites
+
+What things you need to install the software and how to install them
+
+```
+Give examples
+```
+
+### Installing
+
+A step by step series of examples that tell you have to get a development env running
+
+Say what the step will be
+
+```
+Give the example
+```
+
+And repeat
+
+```
+until finished
+```
+
+End with an example of getting some data out of the system or using it for a little demo
+
+## Running the tests
+
+Explain how to run the automated tests for this system
+
+### Break down into end to end tests
+
+Explain what these tests test and why
+
+```
+Give an example
+```
+
+### And coding style tests
+
+Explain what these tests test and why
+
+```
+Give an example
+```
+
+## Deployment
+
+Add additional notes about how to deploy this on a live system
+
+## Built With
+
+* [Dropwizard](http://www.dropwizard.io/1.0.2/docs/) - The web framework used
+* [Maven](https://maven.apache.org/) - Dependency Management
+* [ROME](https://rometools.github.io/rome/) - Used to generate RSS Feeds
+
+## Contributing
+
+Please read [CONTRIBUTING.md](https://gist.github.com/PurpleBooth/b24679402957c63ec426) for details on our code of conduct, and the process for submitting pull requests to us.
+
+## Versioning
+
+We use [SemVer](http://semver.org/) for versioning. For the versions available, see the [tags on this repository](https://github.com/your/project/tags).
+
+## Authors
+
+* **Billie Thompson** - *Initial work* - [PurpleBooth](https://github.com/PurpleBooth)
+
+See also the list of [contributors](https://github.com/your/project/contributors) who participated in this project.
+
+## License
+
+This project is licensed under the MIT License - see the [LICENSE.md](LICENSE.md) file for details
+
+## Acknowledgments
+
+* Hat tip to anyone who's code was used
+* Inspiration
+* etc
+
diff --git a/Code/MonoMutliViewClassifiers/Results/20170922-154253-CMultiV-Benchmark-cq-hist_lss-hist_phog-hist-awaexp-LOG.log b/Code/MonoMutliViewClassifiers/Results/20170922-154253-CMultiV-Benchmark-cq-hist_lss-hist_phog-hist-awaexp-LOG.log
index b724a44a..bbc56280 100644
--- a/Code/MonoMutliViewClassifiers/Results/20170922-154253-CMultiV-Benchmark-cq-hist_lss-hist_phog-hist-awaexp-LOG.log
+++ b/Code/MonoMutliViewClassifiers/Results/20170922-154253-CMultiV-Benchmark-cq-hist_lss-hist_phog-hist-awaexp-LOG.log
@@ -706,3 +706,1269 @@ Classifier configuration :
 2017-09-22 15:50:07,562 DEBUG: Done:	 Determine Train/Test split
 2017-09-22 15:50:07,562 DEBUG: Start:	 RandomSearch best settings with 20 iterations for RandomForest
 2017-09-22 15:50:07,562 DEBUG: Start:	 RandomSearch best settings with 20 iterations for KNN
+2017-09-22 15:50:24,063 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:50:24,063 DEBUG: Start:	 Training
+2017-09-22 15:50:24,385 DEBUG: Done:	 Training
+2017-09-22 15:50:24,385 DEBUG: Start:	 Predicting
+2017-09-22 15:50:24,467 DEBUG: Done:	 Predicting
+2017-09-22 15:50:24,467 DEBUG: Info:	 Time for training and predicting: 16.9445199966[s]
+2017-09-22 15:50:24,467 DEBUG: Start:	 Getting Results
+2017-09-22 15:50:24,495 DEBUG: Done:	 Getting Results
+2017-09-22 15:50:24,495 INFO: Classification on awaexp database for lss-hist with RandomForest, and 5 statistical iterations
+
+accuracy_score on train : 0.998694516971, with STD : 0.0
+accuracy_score on test : 0.677914110429, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- Random Forest with num_esimators : 28, max_depth : 12
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.998694516971
+		- Score on test : 0.677914110429
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.998692810458
+		- Score on test : 0.682779456193
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.998692810458
+		- Score on test : 0.682779456193
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.00130548302872
+		- Score on test : 0.322085889571
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.998694516971
+		- Score on test : 0.677914110429
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.997392433632
+		- Score on test : 0.355995746702
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.672619047619
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.997389033943
+		- Score on test : 0.693251533742
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.998694516971
+		- Score on test : 0.677914110429
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.00130548302872
+		- Score on test : 0.322085889571
+
+
+ Classification took 0:00:16
+2017-09-22 15:50:24,496 INFO: Done:	 Result Analysis
+2017-09-22 15:50:43,708 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:50:43,708 DEBUG: Start:	 Training
+2017-09-22 15:50:43,739 DEBUG: Done:	 Training
+2017-09-22 15:50:43,739 DEBUG: Start:	 Predicting
+2017-09-22 15:50:49,339 DEBUG: Done:	 Predicting
+2017-09-22 15:50:49,340 DEBUG: Info:	 Time for training and predicting: 41.8171880245[s]
+2017-09-22 15:50:49,340 DEBUG: Start:	 Getting Results
+2017-09-22 15:50:49,368 DEBUG: Done:	 Getting Results
+2017-09-22 15:50:49,369 INFO: Classification on awaexp database for lss-hist with KNN, and 5 statistical iterations
+
+accuracy_score on train : 0.711488250653, with STD : 0.0
+accuracy_score on test : 0.59509202454, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- K nearest Neighbors with  n_neighbors: 21
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.711488250653
+		- Score on test : 0.59509202454
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.675477239354
+		- Score on test : 0.551020408163
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.675477239354
+		- Score on test : 0.551020408163
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.288511749347
+		- Score on test : 0.40490797546
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.711488250653
+		- Score on test : 0.59509202454
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.433794414692
+		- Score on test : 0.19395846585
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.771812080537
+		- Score on test : 0.618320610687
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.600522193211
+		- Score on test : 0.496932515337
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.711488250653
+		- Score on test : 0.59509202454
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.288511749347
+		- Score on test : 0.40490797546
+
+
+ Classification took 0:00:41
+2017-09-22 15:50:49,369 INFO: Done:	 Result Analysis
+2017-09-22 15:50:49,472 DEBUG: Start:	 Loading data
+2017-09-22 15:50:49,472 DEBUG: Start:	 Loading data
+2017-09-22 15:50:49,484 DEBUG: Done:	 Loading data
+2017-09-22 15:50:49,484 DEBUG: Done:	 Loading data
+2017-09-22 15:50:49,484 DEBUG: Info:	 Classification - Database:awaexp Feature:lss-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SGD
+2017-09-22 15:50:49,484 DEBUG: Info:	 Classification - Database:awaexp Feature:lss-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMLinear
+2017-09-22 15:50:49,485 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:50:49,485 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:50:49,514 DEBUG: Info:	 Shape X_train:(766, 2000), Length of y_train:766
+2017-09-22 15:50:49,514 DEBUG: Info:	 Shape X_train:(766, 2000), Length of y_train:766
+2017-09-22 15:50:49,515 DEBUG: Info:	 Shape X_test:(326, 2000), Length of y_test:326
+2017-09-22 15:50:49,515 DEBUG: Info:	 Shape X_test:(326, 2000), Length of y_test:326
+2017-09-22 15:50:49,515 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:50:49,515 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:50:49,515 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMLinear
+2017-09-22 15:50:49,515 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SGD
+2017-09-22 15:50:58,244 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:50:58,244 DEBUG: Start:	 Training
+2017-09-22 15:50:58,287 DEBUG: Done:	 Training
+2017-09-22 15:50:58,287 DEBUG: Start:	 Predicting
+2017-09-22 15:50:58,294 DEBUG: Done:	 Predicting
+2017-09-22 15:50:58,295 DEBUG: Info:	 Time for training and predicting: 8.82169294357[s]
+2017-09-22 15:50:58,295 DEBUG: Start:	 Getting Results
+2017-09-22 15:50:58,326 DEBUG: Done:	 Getting Results
+2017-09-22 15:50:58,326 INFO: Classification on awaexp database for lss-hist with SGD, and 5 statistical iterations
+
+accuracy_score on train : 0.955613577023, with STD : 0.0
+accuracy_score on test : 0.791411042945, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SGDClassifier with loss : log, penalty : l2
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.955613577023
+		- Score on test : 0.791411042945
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.956072351421
+		- Score on test : 0.805714285714
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.956072351421
+		- Score on test : 0.805714285714
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0443864229765
+		- Score on test : 0.208588957055
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.955613577023
+		- Score on test : 0.791411042945
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.911426002045
+		- Score on test : 0.589244315989
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.946291560102
+		- Score on test : 0.754010695187
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.966057441253
+		- Score on test : 0.865030674847
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.955613577023
+		- Score on test : 0.791411042945
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0443864229765
+		- Score on test : 0.208588957055
+
+
+ Classification took 0:00:08
+2017-09-22 15:50:58,327 INFO: Done:	 Result Analysis
+2017-09-22 15:51:55,058 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:51:55,058 DEBUG: Start:	 Training
+2017-09-22 15:51:58,941 DEBUG: Done:	 Training
+2017-09-22 15:51:58,941 DEBUG: Start:	 Predicting
+2017-09-22 15:52:00,817 DEBUG: Done:	 Predicting
+2017-09-22 15:52:00,817 DEBUG: Info:	 Time for training and predicting: 71.3445019722[s]
+2017-09-22 15:52:00,818 DEBUG: Start:	 Getting Results
+2017-09-22 15:52:00,845 DEBUG: Done:	 Getting Results
+2017-09-22 15:52:00,845 INFO: Classification on awaexp database for lss-hist with SVMLinear, and 5 statistical iterations
+
+accuracy_score on train : 1.0, with STD : 0.0
+accuracy_score on test : 0.806748466258, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM Linear with C : 3750
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.806748466258
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.809667673716
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.809667673716
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.193251533742
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.806748466258
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.613785770175
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.797619047619
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.822085889571
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.806748466258
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.193251533742
+
+
+ Classification took 0:01:11
+2017-09-22 15:52:00,845 INFO: Done:	 Result Analysis
+2017-09-22 15:52:00,999 DEBUG: Start:	 Loading data
+2017-09-22 15:52:00,999 DEBUG: Start:	 Loading data
+2017-09-22 15:52:01,011 DEBUG: Done:	 Loading data
+2017-09-22 15:52:01,011 DEBUG: Done:	 Loading data
+2017-09-22 15:52:01,011 DEBUG: Info:	 Classification - Database:awaexp Feature:lss-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMRBF
+2017-09-22 15:52:01,011 DEBUG: Info:	 Classification - Database:awaexp Feature:lss-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMPoly
+2017-09-22 15:52:01,011 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:52:01,011 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:52:01,040 DEBUG: Info:	 Shape X_train:(766, 2000), Length of y_train:766
+2017-09-22 15:52:01,040 DEBUG: Info:	 Shape X_train:(766, 2000), Length of y_train:766
+2017-09-22 15:52:01,040 DEBUG: Info:	 Shape X_test:(326, 2000), Length of y_test:326
+2017-09-22 15:52:01,040 DEBUG: Info:	 Shape X_test:(326, 2000), Length of y_test:326
+2017-09-22 15:52:01,041 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:52:01,041 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:52:01,041 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMRBF
+2017-09-22 15:52:01,041 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMPoly
+2017-09-22 15:53:09,464 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:53:09,465 DEBUG: Start:	 Training
+2017-09-22 15:53:14,572 DEBUG: Done:	 Training
+2017-09-22 15:53:14,573 DEBUG: Start:	 Predicting
+2017-09-22 15:53:17,073 DEBUG: Done:	 Predicting
+2017-09-22 15:53:17,074 DEBUG: Info:	 Time for training and predicting: 76.0741391182[s]
+2017-09-22 15:53:17,074 DEBUG: Start:	 Getting Results
+2017-09-22 15:53:17,103 DEBUG: Done:	 Getting Results
+2017-09-22 15:53:17,104 INFO: Classification on awaexp database for lss-hist with SVMPoly, and 5 statistical iterations
+
+accuracy_score on train : 1.0, with STD : 0.0
+accuracy_score on test : 0.800613496933, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM Poly with C : 7894
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.800613496933
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.790996784566
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.790996784566
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.199386503067
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.800613496933
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.603789028059
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.831081081081
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.754601226994
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.800613496933
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.199386503067
+
+
+ Classification took 0:01:16
+2017-09-22 15:53:17,104 INFO: Done:	 Result Analysis
+2017-09-22 15:54:06,512 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:06,513 DEBUG: Start:	 Training
+2017-09-22 15:54:13,323 DEBUG: Done:	 Training
+2017-09-22 15:54:13,324 DEBUG: Start:	 Predicting
+2017-09-22 15:54:17,191 DEBUG: Done:	 Predicting
+2017-09-22 15:54:17,191 DEBUG: Info:	 Time for training and predicting: 136.191291094[s]
+2017-09-22 15:54:17,191 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:17,219 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:17,219 INFO: Classification on awaexp database for lss-hist with SVMRBF, and 5 statistical iterations
+
+accuracy_score on train : 1.0, with STD : 0.0
+accuracy_score on test : 0.530674846626, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : lss-hist	 View shape : (1092, 2000)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM RBF with C : 3750
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.530674846626
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.673773987207
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.673773987207
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.469325153374
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.530674846626
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.127827498141
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.516339869281
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.969325153374
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.530674846626
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.469325153374
+
+
+ Classification took 0:02:16
+2017-09-22 15:54:17,219 INFO: Done:	 Result Analysis
+2017-09-22 15:54:17,347 DEBUG: Start:	 Loading data
+2017-09-22 15:54:17,347 DEBUG: Start:	 Loading data
+2017-09-22 15:54:17,447 DEBUG: Done:	 Loading data
+2017-09-22 15:54:17,447 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : DecisionTree
+2017-09-22 15:54:17,448 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:17,453 DEBUG: Done:	 Loading data
+2017-09-22 15:54:17,453 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : Adaboost
+2017-09-22 15:54:17,454 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:17,460 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:17,461 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:17,461 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:17,461 DEBUG: Start:	 RandomSearch best settings with 20 iterations for DecisionTree
+2017-09-22 15:54:17,466 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:17,466 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:17,467 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:17,467 DEBUG: Start:	 RandomSearch best settings with 20 iterations for Adaboost
+2017-09-22 15:54:27,870 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:27,870 DEBUG: Start:	 Training
+2017-09-22 15:54:27,990 DEBUG: Done:	 Training
+2017-09-22 15:54:27,990 DEBUG: Start:	 Predicting
+2017-09-22 15:54:27,992 DEBUG: Done:	 Predicting
+2017-09-22 15:54:27,992 DEBUG: Info:	 Time for training and predicting: 10.644974947[s]
+2017-09-22 15:54:27,992 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:28,021 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:28,022 INFO: Classification on awaexp database for phog-hist with DecisionTree, and 5 statistical iterations
+
+accuracy_score on train : 0.936031331593, with STD : 0.0
+accuracy_score on test : 0.61963190184, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- Decision Tree with max_depth : 7
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.936031331593
+		- Score on test : 0.61963190184
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.935611038108
+		- Score on test : 0.624242424242
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.935611038108
+		- Score on test : 0.624242424242
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0639686684073
+		- Score on test : 0.38036809816
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.936031331593
+		- Score on test : 0.61963190184
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.872136984892
+		- Score on test : 0.239335879234
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.941798941799
+		- Score on test : 0.616766467066
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.929503916449
+		- Score on test : 0.631901840491
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.936031331593
+		- Score on test : 0.61963190184
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0639686684073
+		- Score on test : 0.38036809816
+
+
+ Classification took 0:00:10
+2017-09-22 15:54:28,022 INFO: Done:	 Result Analysis
+2017-09-22 15:54:29,890 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:29,890 DEBUG: Start:	 Training
+2017-09-22 15:54:30,042 DEBUG: Done:	 Training
+2017-09-22 15:54:30,042 DEBUG: Start:	 Predicting
+2017-09-22 15:54:30,046 DEBUG: Done:	 Predicting
+2017-09-22 15:54:30,046 DEBUG: Info:	 Time for training and predicting: 12.6987230778[s]
+2017-09-22 15:54:30,046 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:30,074 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:30,074 INFO: Classification on awaexp database for phog-hist with Adaboost, and 5 statistical iterations
+
+accuracy_score on train : 1.0, with STD : 0.0
+accuracy_score on test : 0.592024539877, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- Adaboost with num_esimators : 5, base_estimators : DecisionTreeClassifier(class_weight=None, criterion='gini', max_depth=None,
+            max_features=None, max_leaf_nodes=None,
+            min_impurity_split=1e-07, min_samples_leaf=1,
+            min_samples_split=2, min_weight_fraction_leaf=0.0,
+            presort=False, random_state=None, splitter='best')
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.592024539877
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.605341246291
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.605341246291
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.407975460123
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.592024539877
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.184469612979
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.586206896552
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.625766871166
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.592024539877
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.407975460123
+
+
+ Classification took 0:00:12
+2017-09-22 15:54:30,075 INFO: Done:	 Result Analysis
+2017-09-22 15:54:30,216 DEBUG: Start:	 Loading data
+2017-09-22 15:54:30,217 DEBUG: Start:	 Loading data
+2017-09-22 15:54:30,219 DEBUG: Done:	 Loading data
+2017-09-22 15:54:30,219 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : KNN
+2017-09-22 15:54:30,219 DEBUG: Done:	 Loading data
+2017-09-22 15:54:30,219 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:30,220 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : RandomForest
+2017-09-22 15:54:30,220 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:30,236 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:30,236 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:30,236 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:30,236 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:30,236 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:30,236 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:30,236 DEBUG: Start:	 RandomSearch best settings with 20 iterations for KNN
+2017-09-22 15:54:30,236 DEBUG: Start:	 RandomSearch best settings with 20 iterations for RandomForest
+2017-09-22 15:54:35,705 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:35,705 DEBUG: Start:	 Training
+2017-09-22 15:54:35,710 DEBUG: Done:	 Training
+2017-09-22 15:54:35,710 DEBUG: Start:	 Predicting
+2017-09-22 15:54:36,430 DEBUG: Done:	 Predicting
+2017-09-22 15:54:36,431 DEBUG: Info:	 Time for training and predicting: 6.21422100067[s]
+2017-09-22 15:54:36,431 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:36,459 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:36,460 INFO: Classification on awaexp database for phog-hist with KNN, and 5 statistical iterations
+
+accuracy_score on train : 0.648825065274, with STD : 0.0
+accuracy_score on test : 0.58282208589, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- K nearest Neighbors with  n_neighbors: 25
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.648825065274
+		- Score on test : 0.58282208589
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.601481481481
+		- Score on test : 0.549668874172
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.601481481481
+		- Score on test : 0.549668874172
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.351174934726
+		- Score on test : 0.41717791411
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.648825065274
+		- Score on test : 0.58282208589
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.306425042338
+		- Score on test : 0.167469437176
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.695205479452
+		- Score on test : 0.597122302158
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.530026109661
+		- Score on test : 0.509202453988
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.648825065274
+		- Score on test : 0.58282208589
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.351174934726
+		- Score on test : 0.41717791411
+
+
+ Classification took 0:00:06
+2017-09-22 15:54:36,460 INFO: Done:	 Result Analysis
+2017-09-22 15:54:44,234 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:44,234 DEBUG: Start:	 Training
+2017-09-22 15:54:44,511 DEBUG: Done:	 Training
+2017-09-22 15:54:44,511 DEBUG: Start:	 Predicting
+2017-09-22 15:54:44,584 DEBUG: Done:	 Predicting
+2017-09-22 15:54:44,585 DEBUG: Info:	 Time for training and predicting: 14.3676609993[s]
+2017-09-22 15:54:44,585 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:44,617 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:44,617 INFO: Classification on awaexp database for phog-hist with RandomForest, and 5 statistical iterations
+
+accuracy_score on train : 1.0, with STD : 0.0
+accuracy_score on test : 0.680981595092, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- Random Forest with num_esimators : 28, max_depth : 12
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.680981595092
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.679012345679
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.679012345679
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.319018404908
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.680981595092
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.361990440293
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.683229813665
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.674846625767
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 1.0
+		- Score on test : 0.680981595092
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0
+		- Score on test : 0.319018404908
+
+
+ Classification took 0:00:14
+2017-09-22 15:54:44,617 INFO: Done:	 Result Analysis
+2017-09-22 15:54:44,703 DEBUG: Start:	 Loading data
+2017-09-22 15:54:44,703 DEBUG: Start:	 Loading data
+2017-09-22 15:54:44,706 DEBUG: Done:	 Loading data
+2017-09-22 15:54:44,706 DEBUG: Done:	 Loading data
+2017-09-22 15:54:44,706 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SGD
+2017-09-22 15:54:44,706 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMLinear
+2017-09-22 15:54:44,706 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:44,706 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:44,721 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:44,721 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:44,721 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:44,721 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:44,721 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:44,721 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:44,721 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMLinear
+2017-09-22 15:54:44,722 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SGD
+2017-09-22 15:54:46,726 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:46,727 DEBUG: Start:	 Training
+2017-09-22 15:54:46,732 DEBUG: Done:	 Training
+2017-09-22 15:54:46,732 DEBUG: Start:	 Predicting
+2017-09-22 15:54:46,733 DEBUG: Done:	 Predicting
+2017-09-22 15:54:46,734 DEBUG: Info:	 Time for training and predicting: 2.0305621624[s]
+2017-09-22 15:54:46,734 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:46,765 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:46,765 INFO: Classification on awaexp database for phog-hist with SGD, and 5 statistical iterations
+
+accuracy_score on train : 0.592689295039, with STD : 0.0
+accuracy_score on test : 0.564417177914, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SGDClassifier with loss : log, penalty : l2
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.592689295039
+		- Score on test : 0.564417177914
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.347280334728
+		- Score on test : 0.268041237113
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.347280334728
+		- Score on test : 0.268041237113
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.407310704961
+		- Score on test : 0.435582822086
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.592689295039
+		- Score on test : 0.564417177914
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.28121293038
+		- Score on test : 0.219597524388
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.873684210526
+		- Score on test : 0.838709677419
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.216710182768
+		- Score on test : 0.159509202454
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.592689295039
+		- Score on test : 0.564417177914
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.407310704961
+		- Score on test : 0.435582822086
+
+
+ Classification took 0:00:02
+2017-09-22 15:54:46,766 INFO: Done:	 Result Analysis
+2017-09-22 15:54:54,621 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:54:54,621 DEBUG: Start:	 Training
+2017-09-22 15:54:55,155 DEBUG: Done:	 Training
+2017-09-22 15:54:55,155 DEBUG: Start:	 Predicting
+2017-09-22 15:54:55,414 DEBUG: Done:	 Predicting
+2017-09-22 15:54:55,414 DEBUG: Info:	 Time for training and predicting: 10.7104449272[s]
+2017-09-22 15:54:55,414 DEBUG: Start:	 Getting Results
+2017-09-22 15:54:55,443 DEBUG: Done:	 Getting Results
+2017-09-22 15:54:55,443 INFO: Classification on awaexp database for phog-hist with SVMLinear, and 5 statistical iterations
+
+accuracy_score on train : 0.678851174935, with STD : 0.0
+accuracy_score on test : 0.61963190184, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM Linear with C : 3750
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.678851174935
+		- Score on test : 0.61963190184
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.715277777778
+		- Score on test : 0.683673469388
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.715277777778
+		- Score on test : 0.683673469388
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.321148825065
+		- Score on test : 0.38036809816
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.678851174935
+		- Score on test : 0.61963190184
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.370020340933
+		- Score on test : 0.26167425641
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.642411642412
+		- Score on test : 0.585152838428
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.806788511749
+		- Score on test : 0.822085889571
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.678851174935
+		- Score on test : 0.61963190184
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.321148825065
+		- Score on test : 0.38036809816
+
+
+ Classification took 0:00:10
+2017-09-22 15:54:55,443 INFO: Done:	 Result Analysis
+2017-09-22 15:54:55,580 DEBUG: Start:	 Loading data
+2017-09-22 15:54:55,581 DEBUG: Start:	 Loading data
+2017-09-22 15:54:55,582 DEBUG: Done:	 Loading data
+2017-09-22 15:54:55,583 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMPoly
+2017-09-22 15:54:55,583 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:55,583 DEBUG: Done:	 Loading data
+2017-09-22 15:54:55,583 DEBUG: Info:	 Classification - Database:awaexp Feature:phog-hist train_size:0.7, CrossValidation k-folds:5, cores:1, algorithm : SVMRBF
+2017-09-22 15:54:55,584 DEBUG: Start:	 Determine Train/Test split for iteration 1
+2017-09-22 15:54:55,597 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:55,598 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:55,598 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:55,598 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMPoly
+2017-09-22 15:54:55,601 DEBUG: Info:	 Shape X_train:(766, 252), Length of y_train:766
+2017-09-22 15:54:55,601 DEBUG: Info:	 Shape X_test:(326, 252), Length of y_test:326
+2017-09-22 15:54:55,601 DEBUG: Done:	 Determine Train/Test split
+2017-09-22 15:54:55,601 DEBUG: Start:	 RandomSearch best settings with 20 iterations for SVMRBF
+2017-09-22 15:55:08,024 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:55:08,024 DEBUG: Start:	 Training
+2017-09-22 15:55:08,745 DEBUG: Done:	 Training
+2017-09-22 15:55:08,745 DEBUG: Start:	 Predicting
+2017-09-22 15:55:09,099 DEBUG: Done:	 Predicting
+2017-09-22 15:55:09,099 DEBUG: Info:	 Time for training and predicting: 13.5177659988[s]
+2017-09-22 15:55:09,099 DEBUG: Start:	 Getting Results
+2017-09-22 15:55:09,130 DEBUG: Done:	 Getting Results
+2017-09-22 15:55:09,130 INFO: Classification on awaexp database for phog-hist with SVMRBF, and 5 statistical iterations
+
+accuracy_score on train : 0.832898172324, with STD : 0.0
+accuracy_score on test : 0.659509202454, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM RBF with C : 1030
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.832898172324
+		- Score on test : 0.659509202454
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.843902439024
+		- Score on test : 0.689075630252
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.843902439024
+		- Score on test : 0.689075630252
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.167101827676
+		- Score on test : 0.340490797546
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.832898172324
+		- Score on test : 0.659509202454
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.672514283873
+		- Score on test : 0.324949230649
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.79176201373
+		- Score on test : 0.634020618557
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.903394255875
+		- Score on test : 0.754601226994
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.832898172324
+		- Score on test : 0.659509202454
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.167101827676
+		- Score on test : 0.340490797546
+
+
+ Classification took 0:00:13
+2017-09-22 15:55:09,130 INFO: Done:	 Result Analysis
+2017-09-22 15:55:10,542 DEBUG: Done:	 RandomSearch best settings
+2017-09-22 15:55:10,542 DEBUG: Start:	 Training
+2017-09-22 15:55:11,389 DEBUG: Done:	 Training
+2017-09-22 15:55:11,389 DEBUG: Start:	 Predicting
+2017-09-22 15:55:11,870 DEBUG: Done:	 Predicting
+2017-09-22 15:55:11,870 DEBUG: Info:	 Time for training and predicting: 16.2899699211[s]
+2017-09-22 15:55:11,870 DEBUG: Start:	 Getting Results
+2017-09-22 15:55:11,903 DEBUG: Done:	 Getting Results
+2017-09-22 15:55:11,903 INFO: Classification on awaexp database for phog-hist with SVMPoly, and 5 statistical iterations
+
+accuracy_score on train : 0.844647519582, with STD : 0.0
+accuracy_score on test : 0.598159509202, with STD : 0.0
+
+Database configuration : 
+	- Database name : awaexp
+	- View name : phog-hist	 View shape : (1092, 252)
+	- Learning Rate : 0.7
+	- Labels used : 
+	- Number of cross validation folds : 5
+
+Classifier configuration : 
+	- SVM Poly with C : 3750
+	- Executed on 1 core(s) 
+	- Got configuration using randomized search with 20 iterations 
+
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.844647519582
+		- Score on test : 0.598159509202
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.859504132231
+		- Score on test : 0.656167979003
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.859504132231
+		- Score on test : 0.656167979003
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.155352480418
+		- Score on test : 0.401840490798
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.844647519582
+		- Score on test : 0.598159509202
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.705247353516
+		- Score on test : 0.208549836535
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.784482758621
+		- Score on test : 0.573394495413
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.950391644909
+		- Score on test : 0.766871165644
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.844647519582
+		- Score on test : 0.598159509202
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.155352480418
+		- Score on test : 0.401840490798
+
+
+ Classification took 0:00:16
+2017-09-22 15:55:11,904 INFO: Done:	 Result Analysis
+2017-09-22 15:55:12,063 INFO: ### Main Programm for Multiview Classification
+2017-09-22 15:55:12,064 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 15:55:12,064 INFO: ### Main Programm for Multiview Classification
+2017-09-22 15:55:12,064 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 15:55:12,065 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 15:55:12,065 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 15:55:12,065 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 15:55:12,066 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 15:55:12,066 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 15:55:12,066 INFO: Done:	 Read Database Files
+2017-09-22 15:55:12,066 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 15:55:12,067 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 15:55:12,067 INFO: Done:	 Read Database Files
+2017-09-22 15:55:12,067 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 15:55:12,131 INFO: Done:	 Determine validation split
+2017-09-22 15:55:12,131 INFO: Start:	 Determine 5 folds
+2017-09-22 15:55:12,145 INFO: Done:	 Determine validation split
+2017-09-22 15:55:12,145 INFO: Start:	 Determine 5 folds
+2017-09-22 15:58:31,803 INFO: Done:	 Classification
+2017-09-22 15:58:32,358 INFO: Done:	 Classification
+2017-09-22 15:58:32,953 INFO: Done:	 Classification
+2017-09-22 15:58:33,484 INFO: Done:	 Classification
+2017-09-22 15:58:34,069 INFO: Done:	 Classification
+2017-09-22 15:58:34,069 INFO: Info:	 Time for Classification: 202[s]
+2017-09-22 15:58:34,069 INFO: Start:	 Result Analysis for Fusion
+2017-09-22 15:58:34,744 INFO: 		Result for Multiview classification with LateFusion
+
+Average accuracy_score :
+	-On Train : 0.903133159269
+	-On Test : 0.809202453988
+
+Dataset info :
+	-Database name : awaexp
+	-Labels : 
+	-Views : cq-hist, lss-hist, phog-hist
+	-5 folds
+
+Classification configuration : 
+	-Algorithm used : LateFusion with Bayesian Inference using a weight for each view : 0.477239577855, 0.236511491673, 0.286248930471
+	-With monoview classifiers : 
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.903133159269 with STD : 0.0624201116681
+		- Score on test : 0.809202453988 with STD : 0.028140723782
+
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.912705861478 with STD : 0.0474044942861
+		- Score on test : 0.819195169506 with STD : 0.0145004045657
+
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.912705861478 with STD : 0.0474044942861
+		- Score on test : 0.819195169506 with STD : 0.0145004045657
+
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.0968668407311 with STD : 0.0624201116681
+		- Score on test : 0.190797546012 with STD : 0.028140723782
+
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.903133159269 with STD : 0.0624201116681
+		- Score on test : 0.809202453988 with STD : 0.028140723782
+
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.821343523657 with STD : 0.100808821945
+		- Score on test : 0.631495658738 with STD : 0.0389063078408
+
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.869561689481 with STD : 0.0934807591073
+		- Score on test : 0.790289416542 with STD : 0.0635398535929
+
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.969190600522 with STD : 0.0234928916537
+		- Score on test : 0.861349693252 with STD : 0.0677185017254
+
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.903133159269 with STD : 0.0624201116681
+		- Score on test : 0.809202453988 with STD : 0.028140723782
+
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.0968668407311 with STD : 0.0624201116681
+		- Score on test : 0.190797546012 with STD : 0.028140723782
+
+
+2017-09-22 15:58:34,745 INFO: Done:	 Result Analysis
+2017-09-22 15:58:41,403 INFO: Done:	 Classification
+2017-09-22 15:58:41,842 INFO: Done:	 Classification
+2017-09-22 15:58:42,286 INFO: Done:	 Classification
+2017-09-22 15:58:42,730 INFO: Done:	 Classification
+2017-09-22 15:58:43,177 INFO: Done:	 Classification
+2017-09-22 15:58:43,177 INFO: Info:	 Time for Classification: 211[s]
+2017-09-22 15:58:43,177 INFO: Start:	 Result Analysis for Fusion
+2017-09-22 15:58:43,586 INFO: 		Result for Multiview classification with LateFusion
+
+Average accuracy_score :
+	-On Train : 0.682506527415
+	-On Test : 0.669938650307
+
+Dataset info :
+	-Database name : awaexp
+	-Labels : 
+	-Views : cq-hist, lss-hist, phog-hist
+	-5 folds
+
+Classification configuration : 
+	-Algorithm used : LateFusion with Majority Voting 
+	-With monoview classifiers : 
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.682506527415 with STD : 0.00657424454935
+		- Score on test : 0.669938650307 with STD : 0.00688648598793
+
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.674439638676 with STD : 0.0295114256546
+		- Score on test : 0.661890004968 with STD : 0.0327476466189
+
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.674439638676 with STD : 0.0295114256546
+		- Score on test : 0.661890004968 with STD : 0.0327476466189
+
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.317493472585 with STD : 0.00657424454935
+		- Score on test : 0.330061349693 with STD : 0.00688648598793
+
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.682506527415 with STD : 0.00657424454935
+		- Score on test : 0.669938650307 with STD : 0.00688648598793
+
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.369275003784 with STD : 0.0133682817705
+		- Score on test : 0.344948865665 with STD : 0.0143090016406
+
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.693564492842 with STD : 0.0283381941508
+		- Score on test : 0.681019396548 with STD : 0.0334400547567
+
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.664229765013 with STD : 0.0770053525556
+		- Score on test : 0.653987730061 with STD : 0.0853268154636
+
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.682506527415 with STD : 0.00657424454935
+		- Score on test : 0.669938650307 with STD : 0.00688648598793
+
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.317493472585 with STD : 0.00657424454935
+		- Score on test : 0.330061349693 with STD : 0.00688648598793
+
+
+2017-09-22 15:58:43,587 INFO: Done:	 Result Analysis
+2017-09-22 15:58:43,657 INFO: ### Main Programm for Multiview Classification
+2017-09-22 15:58:43,657 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 15:58:43,658 INFO: ### Main Programm for Multiview Classification
+2017-09-22 15:58:43,659 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 15:58:43,659 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 15:58:43,660 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 15:58:43,660 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 15:58:43,662 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 15:58:43,662 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 15:58:43,662 INFO: Done:	 Read Database Files
+2017-09-22 15:58:43,662 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 15:58:43,663 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 15:58:43,663 INFO: Done:	 Read Database Files
+2017-09-22 15:58:43,663 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 15:58:43,755 INFO: Done:	 Determine validation split
+2017-09-22 15:58:43,755 INFO: Done:	 Determine validation split
+2017-09-22 15:58:43,755 INFO: Start:	 Determine 5 folds
+2017-09-22 15:58:43,755 INFO: Start:	 Determine 5 folds
+2017-09-22 16:03:01,278 INFO: Done:	 Classification
+2017-09-22 16:03:02,212 INFO: Done:	 Classification
+2017-09-22 16:03:03,148 INFO: Done:	 Classification
+2017-09-22 16:03:04,087 INFO: Done:	 Classification
+2017-09-22 16:03:05,194 INFO: Done:	 Classification
+2017-09-22 16:03:05,204 INFO: Info:	 Time for Classification: 261[s]
+2017-09-22 16:03:05,205 INFO: Start:	 Result Analysis for Fusion
+2017-09-22 16:03:06,079 INFO: 		Result for Multiview classification with LateFusion
+
+Average accuracy_score :
+	-On Train : 0.827676240209
+	-On Test : 0.760736196319
+
+Dataset info :
+	-Database name : awaexp
+	-Labels : 
+	-Views : cq-hist, lss-hist, phog-hist
+	-5 folds
+
+Classification configuration : 
+	-Algorithm used : LateFusion with SVM for linear 
+	-With monoview classifiers : 
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.827676240209 with STD : 0.0
+		- Score on test : 0.760736196319 with STD : 0.0
+
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.823056300268 with STD : 1.11022302463e-16
+		- Score on test : 0.75625 with STD : 0.0
+
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.823056300268 with STD : 1.11022302463e-16
+		- Score on test : 0.75625 with STD : 0.0
+
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.172323759791 with STD : 0.0
+		- Score on test : 0.239263803681 with STD : 0.0
+
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.827676240209 with STD : 0.0
+		- Score on test : 0.760736196319 with STD : 0.0
+
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.656247838383 with STD : 0.0
+		- Score on test : 0.521826039844 with STD : 0.0
+
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.845730027548 with STD : 0.0
+		- Score on test : 0.770700636943 with STD : 0.0
+
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.801566579634 with STD : 0.0
+		- Score on test : 0.742331288344 with STD : 0.0
+
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.827676240209 with STD : 1.11022302463e-16
+		- Score on test : 0.760736196319 with STD : 0.0
+
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.172323759791 with STD : 0.0
+		- Score on test : 0.239263803681 with STD : 0.0
+
+
+2017-09-22 16:03:06,080 INFO: Done:	 Result Analysis
+2017-09-22 16:04:28,892 INFO: Done:	 Classification
+2017-09-22 16:04:29,728 INFO: Done:	 Classification
+2017-09-22 16:04:30,740 INFO: Done:	 Classification
+2017-09-22 16:04:31,563 INFO: Done:	 Classification
+2017-09-22 16:04:32,313 INFO: Done:	 Classification
+2017-09-22 16:04:32,313 INFO: Info:	 Time for Classification: 348[s]
+2017-09-22 16:04:32,313 INFO: Start:	 Result Analysis for Fusion
+2017-09-22 16:04:32,742 INFO: 		Result for Multiview classification with LateFusion
+
+Average accuracy_score :
+	-On Train : 0.882506527415
+	-On Test : 0.757668711656
+
+Dataset info :
+	-Database name : awaexp
+	-Labels : 
+	-Views : cq-hist, lss-hist, phog-hist
+	-5 folds
+
+Classification configuration : 
+	-Algorithm used : LateFusion with SCM for linear with max_attributes : 14, p : 0.0241145218036 model_type : conjunction has chosen 1 rule(s) 
+	-With monoview classifiers : 
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+		- SGDClassifier with loss : log, penalty : l2
+
+	For Accuracy score using None as sample_weights (higher is better) : 
+		- Score on train : 0.882506527415 with STD : 0.0
+		- Score on test : 0.757668711656 with STD : 0.0
+
+	For F1 score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.867256637168 with STD : 1.11022302463e-16
+		- Score on test : 0.710622710623 with STD : 0.0
+
+	For F-beta score using None as sample_weights, None as labels, 1 as pos_label, binary as average, 1.0 as beta (higher is better) : 
+		- Score on train : 0.867256637168 with STD : 1.11022302463e-16
+		- Score on test : 0.710622710623 with STD : 0.0
+
+	For Hamming loss using None as classes (lower is better) : 
+		- Score on train : 0.117493472585 with STD : 0.0
+		- Score on test : 0.242331288344 with STD : 0.0
+
+	For Jaccard similarity score using None as sample_weights (higher is better) : 
+		- Score on train : 0.882506527415 with STD : 0.0
+		- Score on test : 0.757668711656 with STD : 0.0
+
+	For Matthews correlation coefficient (higher is better) : 
+		- Score on train : 0.78604273629 with STD : 0.0
+		- Score on test : 0.544949260913 with STD : 0.0
+
+	For Precision score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.996610169492 with STD : 0.0
+		- Score on test : 0.881818181818 with STD : 0.0
+
+	For Recall score using None as sample_weights, None as labels, 1 as pos_label, binary as average (higher is better) : 
+		- Score on train : 0.767624020888 with STD : 0.0
+		- Score on test : 0.59509202454 with STD : 0.0
+
+	For ROC AUC score using None as sample_weights, micro as average (higher is better) : 
+		- Score on train : 0.882506527415 with STD : 0.0
+		- Score on test : 0.757668711656 with STD : 0.0
+
+	For Zero one loss using None as sample_weights (lower is better) : 
+		- Score on train : 0.117493472585 with STD : 0.0
+		- Score on test : 0.242331288344 with STD : 2.77555756156e-17
+
+
+2017-09-22 16:04:32,742 INFO: Done:	 Result Analysis
+2017-09-22 16:04:32,827 INFO: ### Main Programm for Multiview Classification
+2017-09-22 16:04:32,828 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 16:04:32,828 INFO: ### Main Programm for Multiview Classification
+2017-09-22 16:04:32,828 INFO: ### Classification - Database : awaexp ; Views : cq-hist, lss-hist, phog-hist ; Algorithm : Fusion ; Cores : 1
+2017-09-22 16:04:32,829 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 16:04:32,830 INFO: Info:	 Shape of cq-hist :(1092, 2688)
+2017-09-22 16:04:32,831 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 16:04:32,831 INFO: Info:	 Shape of lss-hist :(1092, 2000)
+2017-09-22 16:04:32,831 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 16:04:32,832 INFO: Done:	 Read Database Files
+2017-09-22 16:04:32,832 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 16:04:32,832 INFO: Info:	 Shape of phog-hist :(1092, 252)
+2017-09-22 16:04:32,832 INFO: Done:	 Read Database Files
+2017-09-22 16:04:32,832 INFO: Start:	 Determine validation split for ratio 0.7
+2017-09-22 16:04:32,889 INFO: Done:	 Determine validation split
+2017-09-22 16:04:32,889 INFO: Start:	 Determine 5 folds
+2017-09-22 16:04:32,891 INFO: Done:	 Determine validation split
+2017-09-22 16:04:32,891 INFO: Start:	 Determine 5 folds
-- 
GitLab