Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
B
bolsonaro
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Container registry
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Luc Giffon
bolsonaro
Commits
a1a7f767
Commit
a1a7f767
authored
5 years ago
by
Charly Lamothe
Browse files
Options
Downloads
Plain Diff
Merge branch '13-visualization' into 'master'
Resolve "Visualization" Closes
#13
See merge request
!10
parents
4ad6e725
f7c97b65
No related branches found
No related tags found
1 merge request
!10
Resolve "Visualization"
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
code/bolsonaro/models/model_raw_results.py
+1
-1
1 addition, 1 deletion
code/bolsonaro/models/model_raw_results.py
code/compute_results.py
+36
-6
36 additions, 6 deletions
code/compute_results.py
with
37 additions
and
7 deletions
code/bolsonaro/models/model_raw_results.py
+
1
−
1
View file @
a1a7f767
...
...
@@ -25,7 +25,7 @@ class ModelRawResults(object):
@property
def
model_weights
(
self
):
return
self
.
model_weights
return
self
.
_
model_weights
@property
def
training_time
(
self
):
...
...
This diff is collapsed.
Click to expand it.
code/compute_results.py
+
36
−
6
View file @
a1a7f767
...
...
@@ -28,7 +28,6 @@ def extract_scores_across_seeds_and_extracted_forest_sizes(models_dir, results_d
experiment_train_scores
=
dict
()
experiment_dev_scores
=
dict
()
experiment_test_scores
=
dict
()
experiment_weights
=
dict
()
all_extracted_forest_sizes
=
list
()
# Used to check if all losses were computed using the same metric (it should be the case)
...
...
@@ -45,7 +44,6 @@ def extract_scores_across_seeds_and_extracted_forest_sizes(models_dir, results_d
experiment_train_scores
[
seed
]
=
list
()
experiment_dev_scores
[
seed
]
=
list
()
experiment_test_scores
[
seed
]
=
list
()
experiment_weights
[
seed
]
=
list
()
# List the forest sizes in models/{experiment_id}/seeds/{seed}/extracted_forest_sizes
extracted_forest_sizes
=
os
.
listdir
(
extracted_forest_sizes_root_path
)
...
...
@@ -66,8 +64,6 @@ def extract_scores_across_seeds_and_extracted_forest_sizes(models_dir, results_d
experiment_test_scores
[
seed
].
append
(
model_raw_results
.
test_score
)
# Save the metric
experiment_score_metrics
.
append
(
model_raw_results
.
score_metric
)
# Save the weights
#experiment_weights[seed].append(model_raw_results.model_weights)
# Sanity checks
if
len
(
set
(
experiment_score_metrics
))
>
1
:
...
...
@@ -76,7 +72,7 @@ def extract_scores_across_seeds_and_extracted_forest_sizes(models_dir, results_d
raise
ValueError
(
"
The extracted forest sizes aren
'
t the sames across seeds.
"
)
return
experiment_train_scores
,
experiment_dev_scores
,
experiment_test_scores
,
\
all_extracted_forest_sizes
[
0
],
experiment_score_metrics
[
0
]
#, experiment_weights
all_extracted_forest_sizes
[
0
],
experiment_score_metrics
[
0
]
def
extract_scores_across_seeds_and_forest_size
(
models_dir
,
results_dir
,
experiment_id
,
extracted_forest_sizes_number
):
experiment_id_path
=
models_dir
+
os
.
sep
+
str
(
experiment_id
)
# models/{experiment_id}
...
...
@@ -123,6 +119,36 @@ def extract_scores_across_seeds_and_forest_size(models_dir, results_dir, experim
return
experiment_train_scores
,
experiment_dev_scores
,
experiment_test_scores
,
experiment_score_metrics
[
0
]
def
extract_weights_across_seeds
(
models_dir
,
results_dir
,
experiment_id
):
experiment_id_path
=
models_dir
+
os
.
sep
+
str
(
experiment_id
)
# models/{experiment_id}
experiment_seed_root_path
=
experiment_id_path
+
os
.
sep
+
'
seeds
'
# models/{experiment_id}/seeds
experiment_weights
=
dict
()
# For each seed results stored in models/{experiment_id}/seeds
seeds
=
os
.
listdir
(
experiment_seed_root_path
)
seeds
.
sort
(
key
=
int
)
for
seed
in
seeds
:
experiment_seed_path
=
experiment_seed_root_path
+
os
.
sep
+
seed
# models/{experiment_id}/seeds/{seed}
extracted_forest_sizes_root_path
=
experiment_seed_path
+
os
.
sep
+
'
extracted_forest_sizes
'
# models/{experiment_id}/seeds/{seed}/forest_size
# {{seed}:[]}
experiment_weights
[
seed
]
=
list
()
# List the forest sizes in models/{experiment_id}/seeds/{seed}/extracted_forest_sizes
extracted_forest_sizes
=
os
.
listdir
(
extracted_forest_sizes_root_path
)
extracted_forest_sizes
=
[
nb_tree
for
nb_tree
in
extracted_forest_sizes
if
not
'
no_weights
'
in
nb_tree
]
extracted_forest_sizes
.
sort
(
key
=
int
)
for
extracted_forest_size
in
extracted_forest_sizes
:
# models/{experiment_id}/seeds/{seed}/extracted_forest_sizes/{extracted_forest_size}
extracted_forest_size_path
=
extracted_forest_sizes_root_path
+
os
.
sep
+
extracted_forest_size
# Load models/{experiment_id}/seeds/{seed}/extracted_forest_sizes/{extracted_forest_size}/model_raw_results.pickle file
model_raw_results
=
ModelRawResults
.
load
(
extracted_forest_size_path
)
# Save the weights
experiment_weights
[
seed
].
append
(
model_raw_results
.
model_weights
)
return
experiment_weights
if
__name__
==
"
__main__
"
:
# get environment variables in .env
load_dotenv
(
find_dotenv
(
'
.env
'
))
...
...
@@ -366,7 +392,7 @@ if __name__ == "__main__":
omp_with_params_experiment_score_metric
=
extract_scores_across_seeds_and_extracted_forest_sizes
(
args
.
models_dir
,
args
.
results_dir
,
int
(
args
.
experiment_ids
[
2
]))
#omp_with_params_without_weights
logger
.
info
(
'
Loading omp_
no_
weights experiment scores...
'
)
logger
.
info
(
'
Loading omp_
with_params without
weights experiment scores...
'
)
omp_with_params_without_weights_train_scores
,
omp_with_params_without_weights_dev_scores
,
omp_with_params_without_weights_test_scores
,
_
,
\
omp_with_params_experiment_score_metric
=
extract_scores_across_seeds_and_extracted_forest_sizes
(
args
.
models_dir
,
args
.
results_dir
,
int
(
args
.
experiment_ids
[
2
]),
weights
=
False
)
...
...
@@ -401,6 +427,10 @@ if __name__ == "__main__":
xlabel
=
'
Number of trees extracted
'
,
ylabel
=
experiments_score_metric
,
title
=
'
Loss values of {}
\n
using best params of previous stages
'
.
format
(
args
.
dataset_name
))
experiment_weights
=
extract_weights_across_seeds
(
args
.
models_dir
,
args
.
results_dir
,
args
.
experiment_ids
[
2
])
Plotter
.
weight_density
(
experiment_weights
,
os
.
path
.
join
(
output_path
,
'
weight_density.png
'
))
elif
args
.
stage
==
5
:
# Retreive the extracted forest sizes number used in order to have a base forest axis as long as necessary
extracted_forest_sizes_number
=
retreive_extracted_forest_sizes_number
(
args
.
models_dir
,
int
(
args
.
experiment_ids
[
1
]))
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment