-
Notifications
You must be signed in to change notification settings - Fork 20
/
configs.py
167 lines (134 loc) · 6.17 KB
/
configs.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
"""FocusSeq2Seq
Copyright (c) 2019-present NAVER Corp.
MIT license
"""
import argparse
import pprint
import yaml
def str2bool(v):
if v.lower() in ('yes', 'true', 't', 'y', '1'):
return True
elif v.lower() in ('no', 'false', 'f', 'n', '0'):
return False
else:
raise argparse.ArgumentTypeError('Boolean value expected.')
class Config(object):
def __init__(self, **kwargs):
"""Configuration Class: set kwargs as class attributes with setattr"""
for k, v in kwargs.items():
setattr(self, k, v)
@property
def config_str(self):
return pprint.pformat(self.__dict__)
def __repr__(self):
"""Pretty-print configurations in alphabetical order"""
config_str = 'Configurations\n'
config_str += self.config_str
return config_str
def save(self, path):
with open(path, 'w') as f:
yaml.dump(self.__dict__, f, default_flow_style=False)
@classmethod
def load(cls, path):
with open(path, 'r') as f:
kwargs = yaml.load(f)
return Config(**kwargs)
def read_config(path):
return Config.load(path)
def get_config(parse=True, **optional_kwargs):
"""
Get configurations as attributes of class
1. Parse configurations with argparse.
2. Create Config class initilized with parsed kwargs.
3. Return Config class.
"""
parser = argparse.ArgumentParser()
# Task / Model / Data
parser.add_argument('--task', type=str, default='QG',
choices=['QG', 'SM'],
help='QG: Question Generation / SM: Summarization')
parser.add_argument('--model', type=str, default='NQG',
choices=['NQG', 'PG'],
help='NQG: NQG++ (Zhou et al. 2017) / PG: Pointer Generator (See et al. 2017)')
parser.add_argument('--data', type=str, default='squad',
choices=['squad', 'cnndm'])
# Training
parser.add_argument('--epochs', type=int, default=20,
help='num_epochs')
parser.add_argument('--batch_size', type=int, default=64,
help='batch size')
parser.add_argument('--lr', type=float, default=0.001,
help='learning rate')
parser.add_argument('--clip', type=float, default=5.0,
help='gradient clip norm')
parser.add_argument('--no_clip', action='store_true',
help="Not to use gradient clipping")
parser.add_argument('--optim', type=str, default='adam',
choices=['adam', 'amsgrad', 'adagrad'],
help='optimizer')
parser.add_argument('--dropout', type=float, default=0.0)
parser.add_argument('--dry', action='store_true',
help='Run training script without actually running training steps. Debugging only')
parser.add_argument('--seed', type=int, default=123,
help='Random seed')
# Evaluation
parser.add_argument('--eval_only', action='store_true')
parser.add_argument('--load_ckpt', type=int, default=9)
parser.add_argument('--eval_batch_size', type=int, default=32,
help='batch size during evaluation')
parser.add_argument('--val_data_size', type=int, default=1000,
help='number of examples for validation / Use (for debugging) when evaluation for summarization takes so long')
# Seq2Seq Model
parser.add_argument('--vocab_size', type=int, default=20000)
parser.add_argument('--embed_size', type=int, default=300)
parser.add_argument('--enc_hidden_size', type=int, default=512)
parser.add_argument('--dec_hidden_size', type=int, default=256)
parser.add_argument('--num_layers', type=int, default=1)
parser.add_argument('--rnn', type=str, default='GRU')
parser.add_argument('--weight_tie', type=str2bool, default=True,
help='output layer tied with embedding')
parser.add_argument('--embedding_freeze', type=str2bool, default=False,
help='Freeze word embedding during training')
parser.add_argument('--load_glove', type=str2bool, default=True,
help='Initialize word embedding from glove (NQG++ only)')
parser.add_argument('--feature_rich', action='store_true',
help='Use linguistic features (POS/NER/Word Case/Answer position; NQG++ only)')
parser.add_argument('--coverage_lambda', type=float, default=1.0,
help='hyperparameter for coverage (Pointer Generator only)')
# Seq2Seq Decoding
parser.add_argument('--decoding', type=str, default='beam',
choices=['greedy', 'beam', 'diverse_beam', 'topk_sampling'])
parser.add_argument('--beam_k', type=int, default=1)
parser.add_argument('--temperature', type=float, default=1.0)
parser.add_argument('--diversity_lambda', type=float, default=0.5)
parser.add_argument('--decode_k', type=int, default=1)
parser.add_argument('--mixture_decoder', action='store_true',
help='Hard Uniform Mixture Decoder (Shen et al. 2018)')
# Focus
parser.add_argument('--use_focus', type=str2bool, default=True,
help='whether to use focus or not')
parser.add_argument('--eval_focus_oracle', action='store_true',
help='Feed focus guide even during test time')
# Selector
parser.add_argument('--threshold', type=float, default=0.15,
help='focus binarization threshold')
# Mixture
parser.add_argument('--n_mixture', type=int, default=1,
help='Number of mixtures for Selector (Ours) or Mixture Decoder (Shen et al. 2018)')
if parse:
kwargs = parser.parse_args()
else:
kwargs = parser.parse_known_args()[0]
# Namespace => Dictionary
kwargs = vars(kwargs)
kwargs.update(optional_kwargs)
return Config(**kwargs)
if __name__ == '__main__':
config = get_config()
# Save
config.save('config.txt')
# Load
loaded_config = read_config('config.txt')
assert config.__dict__ == loaded_config.__dict__
import os
os.remove('config.txt')