Skip to content
Snippets Groups Projects
Select Git revision
  • master default
  • object
  • develop protected
  • private_algos
  • cuisine
  • SMOTE
  • revert-76c4cca5
  • archive protected
  • no_graphviz
9 results

summit

  • Clone with SSH
  • Clone with HTTPS
  • Pipeline status License: GPL v3 Coverage

    Supervised MultiModal Integration Tool's Readme

    This project aims to be an easy-to-use solution to run a prior benchmark on a dataset and evaluate mono- & multi-view algorithms capacity to classify it correctly.

    Getting Started

    SuMMIT has been designed and uses continuous integration for Linux platforms (ubuntu 18.04), but we try to keep it as compatible as possible with Mac and Windows.

    Platform Last positive test
    Linux Pipeline status
    Mac 1st of May, 2020
    Windows 1st of May, 2020

    Prerequisites

    To be able to use this project, you'll need :

    And the following python modules will be automatically installed :

    • numpy, scipy,
    • matplotlib - Used to plot results,
    • sklearn - Used for the monoview classifiers,
    • joblib - Used to compute on multiple threads,
    • h5py - Used to generate HDF5 datasets on hard drive and use them to spare RAM,
    • pickle - Used to store some results,
    • pandas - Used to manipulate data efficiently,
    • six -
    • m2r - Used to generate documentation from the readme,
    • docutils - Used to generate documentation,
    • pyyaml - Used to read the config files,
    • plotly - Used to generate interactive HTML visuals,
    • tabulate - Used to generated the confusion matrix.

    Installing

    Once you cloned the project from the gitlab repository, you just have to use :

    cd path/to/summit/
    pip install -e .

    In the summit directory to install SuMMIT and its dependencies.

    Running the tests

    To run the test suite of SuMMIT, run :

    cd path/to/summit
    pip install -e .[dev]
    pytest

    The coverage report is automatically generated and stored in the htmlcov/ directory

    Building the documentation

    To locally build the documentation run :

    cd path/to/summit
    pip install -e .[doc]
    python setup.py build_sphinx

    The built html files will be stored in path/to/summit/build/sphinx/html

    Running on simulated data

    For your first go with SuMMIT, you can run it on simulated data with

    python
    >>> from summit.execute import execute
    >>> execute("example 1")

    This will run the benchmark of documentation's Example 1.

    For more information about the examples, see the documentation. Results will, by default, be stored in the results directory of the installation path : path/to/summit/multiview_platform/examples/results.

    The documentation proposes a detailed interpretation of the results and arguments of SuMMIT through 6 tutorials.

    Dataset compatibility

    In order to start a benchmark on your own dataset, you need to format it so SuMMIT can use it. To do so, a python script is provided.

    For more information, see Example 5

    Running on your dataset

    Once you have formatted your dataset, to run SuMMIT on it you need to modify the config file as

    name: ["your_file_name"]
    pathf: "path/to/your/dataset"

    It is however highly recommended to follow the documentation's tutorials to learn the use of each parameter.

    Authors

    • Baptiste BAUVIN
    • Dominique BENIELLI
    • Alexis PROD'HOMME