forked from sumitkeswani/Captioning-VQA
-
Notifications
You must be signed in to change notification settings - Fork 0
/
VQA_config.py
25 lines (22 loc) · 1.02 KB
/
VQA_config.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
# paths
qa_path = 'vqa' # directory containing the question and annotation jsons
train_path = 'mscoco/train2014' # directory of training images
val_path = 'mscoco/val2014' # directory of validation images
test_path = 'mscoco/test2015' # directory of test images
preprocessed_path = '/ssd/resnet-14x14.h5' # path where preprocessed features are saved to and loaded from
vocabulary_path = 'vocab.json' # path where the used vocabularies for question and answers are saved to
task = 'OpenEnded'
dataset = 'mscoco'
# preprocess config
preprocess_batch_size = 64
image_size = 448 # scale shorter end of image to this size and centre crop
output_size = image_size // 32 # size of the feature maps after processing through a network
output_features = 2048 # number of feature maps thereof
central_fraction = 0.875 # only take this much of the centre when scaling and centre cropping
# training config
epochs = 50
batch_size = 128
initial_lr = 1e-3 # default Adam lr
lr_halflife = 50000 # in iterations
data_workers = 8
max_answers = 3000