Skip to content

Commit

Permalink
instance segmentation SOTA params
Browse files Browse the repository at this point in the history
  • Loading branch information
JSabadin committed Jan 29, 2025
1 parent f04678f commit 6d27f76
Show file tree
Hide file tree
Showing 6 changed files with 33 additions and 15 deletions.
1 change: 1 addition & 0 deletions configs/detection_heavy_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@ loader:
dataset_name: coco_test

trainer:
precision: "16-mixed"
preprocessing:
train_image_size: [384, 512]
keep_aspect_ratio: true
Expand Down
1 change: 1 addition & 0 deletions configs/detection_light_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@ loader:
dataset_name: coco_test

trainer:
precision: "16-mixed"
preprocessing:
train_image_size: [384, 512]
keep_aspect_ratio: true
Expand Down
16 changes: 12 additions & 4 deletions configs/instance_segmentation_heavy_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@ loader:
dataset_name: coco_test

trainer:
precision: "16-mixed"
preprocessing:
train_image_size: [384, 512]
keep_aspect_ratio: true
Expand All @@ -27,10 +28,10 @@ trainer:

batch_size: 8
epochs: &epochs 300
accumulate_grad_batches: 8 # For best results, always accumulate gradients to effectively use 64 batch size
n_workers: 8
n_workers: 4
validation_interval: 10
n_log_images: 8
gradient_clip_val: 10

callbacks:
- name: EMACallback
Expand All @@ -40,15 +41,22 @@ trainer:
decay_tau: 2000
- name: ExportOnTrainEnd
- name: TestOnTrainEnd
- name: GradientAccumulationScheduler
params:
scheduling: # warmup phase is 3 epochs
0: 1
1: 4
2: 8 # For best results, always accumulate gradients to effectively use 64 batch size

training_strategy:
name: "TripleLRSGDStrategy"
params:
warmup_epochs: 3
warmup_bias_lr: 0.1
warmup_bias_lr: 0.0
warmup_momentum: 0.8
lr: 0.01
lre: 0.0001
momentum: 0.937
weight_decay: 0.0005
nesterov: True
nesterov: True
cosine_annealing: False
18 changes: 13 additions & 5 deletions configs/instance_segmentation_light_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,8 +16,9 @@ loader:
dataset_name: coco_test

trainer:
precision: "16-mixed"
preprocessing:
train_image_size: [384, 512]
train_image_size: [384, 384]
keep_aspect_ratio: true
normalize:
active: true
Expand All @@ -27,10 +28,10 @@ trainer:

batch_size: 8
epochs: &epochs 300
accumulate_grad_batches: 8 # For best results, always accumulate gradients to effectively use 64 batch size
n_workers: 8
n_workers: 4
validation_interval: 10
n_log_images: 8
gradient_clip_val: 10

callbacks:
- name: EMACallback
Expand All @@ -40,15 +41,22 @@ trainer:
decay_tau: 2000
- name: ExportOnTrainEnd
- name: TestOnTrainEnd
- name: GradientAccumulationScheduler
params:
scheduling: # warmup phase is 3 epochs
0: 1
1: 4
2: 8 # For best results, always accumulate gradients to effectively use 64 batch size

training_strategy:
name: "TripleLRSGDStrategy"
params:
warmup_epochs: 3
warmup_bias_lr: 0.1
warmup_bias_lr: 0.0
warmup_momentum: 0.8
lr: 0.01
lre: 0.0001
momentum: 0.937
weight_decay: 0.0005
nesterov: True
nesterov: True
cosine_annealing: False
6 changes: 3 additions & 3 deletions configs/keypoint_bbox_heavy_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ trainer:
std: [1, 1, 1]

batch_size: 8
epochs: &epochs 100
epochs: &epochs 300
n_workers: 4
validation_interval: 10
n_log_images: 8
Expand All @@ -46,10 +46,10 @@ trainer:
- name: TestOnTrainEnd
- name: GradientAccumulationScheduler
params:
scheduling:
scheduling: # warmup phase is 3 epochs
0: 1
1: 4
2: 8
2: 8 # For best results, always accumulate gradients to effectively use 64 batch size

training_strategy:
name: "TripleLRSGDStrategy"
Expand Down
6 changes: 3 additions & 3 deletions configs/keypoint_bbox_light_model.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ trainer:
std: [1, 1, 1]

batch_size: 8
epochs: &epochs 100
epochs: &epochs 300
n_workers: 4
validation_interval: 10
n_log_images: 8
Expand All @@ -46,10 +46,10 @@ trainer:
- name: TestOnTrainEnd
- name: GradientAccumulationScheduler
params:
scheduling:
scheduling: # warmup phase is 3 epochs
0: 1
1: 4
2: 8
2: 8 # For best results, always accumulate gradients to effectively use 64 batch size

training_strategy:
name: "TripleLRSGDStrategy"
Expand Down

0 comments on commit 6d27f76

Please sign in to comment.