-
Notifications
You must be signed in to change notification settings - Fork 122
Loading model #168
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
shakibamrd
wants to merge
13
commits into
Develop_copy
Choose a base branch
from
loading_model
base: Develop_copy
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Loading model #168
Changes from all commits
Commits
Show all changes
13 commits
Select commit
Hold shift + click to select a range
01f97f0
added config to reproduce drnas results
shakibamrd 6cb1e7b
update config for drnas
shakibamrd 788b517
add batch_size and train_portion for eval mode
shakibamrd d021f05
add todo to drnas optimizer
shakibamrd f12c73d
fixed merge conflict with remote branch
shakibamrd 73c70c8
remove auxiliry head when running nb301
shakibamrd 44891c9
bug fix: able to do full eval with spaces
shakibamrd 9766329
refactored naslib runner
shakibamrd cf68cef
added feature to handels the number of classes based on the dataset
shakibamrd 7ba12ea
added arch_weights as additional checkpointables
shakibamrd f0aa6a8
added set_checkpointables to all optimizers
shakibamrd 09641d0
bug fix, updated order of before_training and _set_checkpoint for eva…
shakibamrd a17a284
resolve comments made on the pull request
shakibamrd File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,125 @@ | ||
# options cifar10, cifar100, ImageNet16-120 reports on their test acc is avaliable | ||
dataset: ImageNet16-120 | ||
# in the code base the deafult value for the seed is 2. | ||
# using random seeds that are logged but log files are not provided | ||
# not mentioned in the paper what are the random seeds are | ||
seed: 99 | ||
# darts (or nb301) | ||
# nb201 | ||
search_space: nasbench301 | ||
out_dir: run | ||
optimizer: drnas | ||
|
||
search: | ||
checkpoint_freq: 5 | ||
# default value batch size in code is 64 | ||
batch_size: 64 | ||
# lr_rate for progressive and original: 0.025 | ||
learning_rate: 0.025 | ||
# lr_rate for progressive and original: 0.025 | ||
learning_rate_min: 0.001 | ||
momentum: 0.9 | ||
# weight_decay for progressive and original: 0.0003 | ||
weight_decay: 0.0003 | ||
# for cifar10 the learning process is 2 stages of 25 epochs each | ||
# in code it states that the number of training epochs has the default value of 100 in nb201 | ||
epochs: 100 | ||
warm_start_epochs: 0 | ||
grad_clip: 5 | ||
# for cifar10 the train and optimization data (50k) is equally partitioned | ||
train_portion: 0.5 | ||
# for cifar10 the train and optimization data (50k) is equally partitioned | ||
data_size: 25000 | ||
|
||
# for the four args the values are same for oridinary and progressive mode for nb201 | ||
cutout: False | ||
cutout_length: 16 | ||
cutout_prob: 1.0 | ||
drop_path_prob: 0.0 | ||
|
||
# for nb201 this value is false | ||
unrolled: False | ||
arch_learning_rate: 0.0003 | ||
# not mentiond for progressive mode but for ordinary it is 1e-3 in nb201 | ||
arch_weight_decay: 0.001 | ||
output_weights: True | ||
|
||
fidelity: 200 | ||
|
||
# GDAS | ||
tau_max: 10 | ||
tau_min: 0.1 | ||
|
||
# RE | ||
sample_size: 10 | ||
population_size: 100 | ||
|
||
#LS | ||
num_init: 10 | ||
|
||
#GSparsity-> Uncomment the lines below for GSparsity | ||
#seed: 50 | ||
#grad_clip: 0 | ||
#threshold: 0.000001 | ||
#weight_decay: 120 | ||
#learning_rate: 0.01 | ||
#momentum: 0.8 | ||
#normalization: div | ||
#normalization_exponent: 0.5 | ||
#batch_size: 256 | ||
#learning_rate_min: 0.0001 | ||
#epochs: 100 | ||
#warm_start_epochs: 0 | ||
#train_portion: 0.9 | ||
#data_size: 25000 | ||
|
||
|
||
# BANANAS | ||
k: 10 | ||
num_ensemble: 3 | ||
acq_fn_type: its | ||
acq_fn_optimization: mutation | ||
encoding_type: path | ||
num_arches_to_mutate: 2 | ||
max_mutations: 1 | ||
num_candidates: 100 | ||
|
||
# BasePredictor | ||
predictor_type: var_sparse_gp | ||
debug_predictor: False | ||
|
||
evaluation: | ||
checkpoint_freq: 30 | ||
# Neither the paper nor the code base indicates the batch size but the default value is 64 | ||
batch_size: 64 | ||
|
||
learning_rate: 0.025 | ||
learning_rate_min: 0.00 | ||
# momentum is 0.9 | ||
momentum: 0.9 | ||
# for cifar weight_decay is 3e-4 | ||
weight_decay: 0.0003 | ||
# cifar's eval is 600 epochs, for imagenet it is 250 | ||
epochs: 250 | ||
# for image net it has 5 epochs of warm starting | ||
warm_start_epochs: 5 | ||
grad_clip: 5 | ||
# uses the whole training data of cifar10 (50K) to train from scratch for 600 epochs | ||
train_portion: 1. | ||
data_size: 50000 | ||
|
||
# cifar10 the cutout is done to have fair comparisons with previous work | ||
cutout: True | ||
# cifar10 cutout length is 16 | ||
cutout_length: 16 | ||
# cifar10 the cutout is done to have fair comparisons with previous work | ||
cutout_prob: 1.0 | ||
# cifar drop out is 0.3 | ||
drop_path_prob: 0.2 | ||
# cifar auxiliary is 0.4 | ||
auxiliary_weight: 0.4 | ||
|
||
|
||
|
||
# has a partial channel variable that for oridinary is 1 and in progressive mode has 4 as the default value. | ||
# mentions some things about regularization scale of l2 and kl (used for dirichlet) in code of nb201 |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -138,6 +138,7 @@ def single_evaluate(self, test_data, zc_api): | |
logger.info("Querying the predictor") | ||
query_time_start = time.time() | ||
|
||
# TODO: shouldn't mode="val" be passed? | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Makes sense for me. |
||
_, _, test_loader, _, _ = utils.get_train_val_loaders(self.config) | ||
|
||
# Iterate over the architectures, instantiate a graph with each architecture | ||
|
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -318,3 +318,4 @@ def get_arch_as_string(self, arch): | |
else: | ||
str_arch = str(arch) | ||
return str_arch | ||
|
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In order to make yaml files generally more readable, should focus only on specific optimizer settings @Neonkraft ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Agreed.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The darts_defualts.yaml was reverted to the format of the Develop_copy branch.