Skip to content

Commit 9b3d636

Browse files
update sam2.1 yaml
1 parent 50b26d0 commit 9b3d636

4 files changed

Lines changed: 56 additions & 56 deletions

File tree

projects/sam2/configs/sam2.1/sam2.1_hiera_b+.yaml

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2,18 +2,18 @@
22

33
# Model
44
model:
5-
_target_: sam2.modeling.sam2_base.SAM2Base
5+
_target_: projects.sam2.modeling.sam2_base.SAM2Base
66
image_encoder:
7-
_target_: sam2.modeling.backbones.image_encoder.ImageEncoder
7+
_target_: projects.sam2.modeling.backbones.image_encoder.ImageEncoder
88
scalp: 1
99
trunk:
10-
_target_: sam2.modeling.backbones.hieradet.Hiera
10+
_target_: projects.sam2.modeling.backbones.hieradet.Hiera
1111
embed_dim: 112
1212
num_heads: 2
1313
neck:
14-
_target_: sam2.modeling.backbones.image_encoder.FpnNeck
14+
_target_: projects.sam2.modeling.backbones.image_encoder.FpnNeck
1515
position_encoding:
16-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
16+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
1717
num_pos_feats: 256
1818
normalize: true
1919
scale: null
@@ -24,17 +24,17 @@ model:
2424
fpn_interp_model: nearest
2525

2626
memory_attention:
27-
_target_: sam2.modeling.memory_attention.MemoryAttention
27+
_target_: projects.sam2.modeling.memory_attention.MemoryAttention
2828
d_model: 256
2929
pos_enc_at_input: true
3030
layer:
31-
_target_: sam2.modeling.memory_attention.MemoryAttentionLayer
31+
_target_: projects.sam2.modeling.memory_attention.MemoryAttentionLayer
3232
activation: relu
3333
dim_feedforward: 2048
3434
dropout: 0.1
3535
pos_enc_at_attn: false
3636
self_attention:
37-
_target_: sam2.modeling.sam.transformer.RoPEAttention
37+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
3838
rope_theta: 10000.0
3939
feat_sizes: [64, 64]
4040
embedding_dim: 256
@@ -45,7 +45,7 @@ model:
4545
pos_enc_at_cross_attn_keys: true
4646
pos_enc_at_cross_attn_queries: false
4747
cross_attention:
48-
_target_: sam2.modeling.sam.transformer.RoPEAttention
48+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
4949
rope_theta: 10000.0
5050
feat_sizes: [64, 64]
5151
rope_k_repeat: True
@@ -57,23 +57,23 @@ model:
5757
num_layers: 4
5858

5959
memory_encoder:
60-
_target_: sam2.modeling.memory_encoder.MemoryEncoder
60+
_target_: projects.sam2.modeling.memory_encoder.MemoryEncoder
6161
out_dim: 64
6262
position_encoding:
63-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
63+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
6464
num_pos_feats: 64
6565
normalize: true
6666
scale: null
6767
temperature: 10000
6868
mask_downsampler:
69-
_target_: sam2.modeling.memory_encoder.MaskDownSampler
69+
_target_: projects.sam2.modeling.memory_encoder.MaskDownSampler
7070
kernel_size: 3
7171
stride: 2
7272
padding: 1
7373
fuser:
74-
_target_: sam2.modeling.memory_encoder.Fuser
74+
_target_: projects.sam2.modeling.memory_encoder.Fuser
7575
layer:
76-
_target_: sam2.modeling.memory_encoder.CXBlock
76+
_target_: projects.sam2.modeling.memory_encoder.CXBlock
7777
dim: 256
7878
kernel_size: 7
7979
padding: 3

projects/sam2/configs/sam2.1/sam2.1_hiera_l.yaml

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2,22 +2,22 @@
22

33
# Model
44
model:
5-
_target_: sam2.modeling.sam2_base.SAM2Base
5+
_target_: projects.sam2.modeling.sam2_base.SAM2Base
66
image_encoder:
7-
_target_: sam2.modeling.backbones.image_encoder.ImageEncoder
7+
_target_: projects.sam2.modeling.backbones.image_encoder.ImageEncoder
88
scalp: 1
99
trunk:
10-
_target_: sam2.modeling.backbones.hieradet.Hiera
10+
_target_: projects.sam2.modeling.backbones.hieradet.Hiera
1111
embed_dim: 144
1212
num_heads: 2
1313
stages: [2, 6, 36, 4]
1414
global_att_blocks: [23, 33, 43]
1515
window_pos_embed_bkg_spatial_size: [7, 7]
1616
window_spec: [8, 4, 16, 8]
1717
neck:
18-
_target_: sam2.modeling.backbones.image_encoder.FpnNeck
18+
_target_: projects.sam2.modeling.backbones.image_encoder.FpnNeck
1919
position_encoding:
20-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
20+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
2121
num_pos_feats: 256
2222
normalize: true
2323
scale: null
@@ -28,17 +28,17 @@ model:
2828
fpn_interp_model: nearest
2929

3030
memory_attention:
31-
_target_: sam2.modeling.memory_attention.MemoryAttention
31+
_target_: projects.sam2.modeling.memory_attention.MemoryAttention
3232
d_model: 256
3333
pos_enc_at_input: true
3434
layer:
35-
_target_: sam2.modeling.memory_attention.MemoryAttentionLayer
35+
_target_: projects.sam2.modeling.memory_attention.MemoryAttentionLayer
3636
activation: relu
3737
dim_feedforward: 2048
3838
dropout: 0.1
3939
pos_enc_at_attn: false
4040
self_attention:
41-
_target_: sam2.modeling.sam.transformer.RoPEAttention
41+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
4242
rope_theta: 10000.0
4343
feat_sizes: [64, 64]
4444
embedding_dim: 256
@@ -49,7 +49,7 @@ model:
4949
pos_enc_at_cross_attn_keys: true
5050
pos_enc_at_cross_attn_queries: false
5151
cross_attention:
52-
_target_: sam2.modeling.sam.transformer.RoPEAttention
52+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
5353
rope_theta: 10000.0
5454
feat_sizes: [64, 64]
5555
rope_k_repeat: True
@@ -61,23 +61,23 @@ model:
6161
num_layers: 4
6262

6363
memory_encoder:
64-
_target_: sam2.modeling.memory_encoder.MemoryEncoder
64+
_target_: projects.sam2.modeling.memory_encoder.MemoryEncoder
6565
out_dim: 64
6666
position_encoding:
67-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
67+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
6868
num_pos_feats: 64
6969
normalize: true
7070
scale: null
7171
temperature: 10000
7272
mask_downsampler:
73-
_target_: sam2.modeling.memory_encoder.MaskDownSampler
73+
_target_: projects.sam2.modeling.memory_encoder.MaskDownSampler
7474
kernel_size: 3
7575
stride: 2
7676
padding: 1
7777
fuser:
78-
_target_: sam2.modeling.memory_encoder.Fuser
78+
_target_: projects.sam2.modeling.memory_encoder.Fuser
7979
layer:
80-
_target_: sam2.modeling.memory_encoder.CXBlock
80+
_target_: projects.sam2.modeling.memory_encoder.CXBlock
8181
dim: 256
8282
kernel_size: 7
8383
padding: 3

projects/sam2/configs/sam2.1/sam2.1_hiera_s.yaml

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2,21 +2,21 @@
22

33
# Model
44
model:
5-
_target_: sam2.modeling.sam2_base.SAM2Base
5+
_target_: projects.sam2.modeling.sam2_base.SAM2Base
66
image_encoder:
7-
_target_: sam2.modeling.backbones.image_encoder.ImageEncoder
7+
_target_: projects.sam2.modeling.backbones.image_encoder.ImageEncoder
88
scalp: 1
99
trunk:
10-
_target_: sam2.modeling.backbones.hieradet.Hiera
10+
_target_: projects.sam2.modeling.backbones.hieradet.Hiera
1111
embed_dim: 96
1212
num_heads: 1
1313
stages: [1, 2, 11, 2]
1414
global_att_blocks: [7, 10, 13]
1515
window_pos_embed_bkg_spatial_size: [7, 7]
1616
neck:
17-
_target_: sam2.modeling.backbones.image_encoder.FpnNeck
17+
_target_: projects.sam2.modeling.backbones.image_encoder.FpnNeck
1818
position_encoding:
19-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
19+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
2020
num_pos_feats: 256
2121
normalize: true
2222
scale: null
@@ -27,17 +27,17 @@ model:
2727
fpn_interp_model: nearest
2828

2929
memory_attention:
30-
_target_: sam2.modeling.memory_attention.MemoryAttention
30+
_target_: projects.sam2.modeling.memory_attention.MemoryAttention
3131
d_model: 256
3232
pos_enc_at_input: true
3333
layer:
34-
_target_: sam2.modeling.memory_attention.MemoryAttentionLayer
34+
_target_: projects.sam2.modeling.memory_attention.MemoryAttentionLayer
3535
activation: relu
3636
dim_feedforward: 2048
3737
dropout: 0.1
3838
pos_enc_at_attn: false
3939
self_attention:
40-
_target_: sam2.modeling.sam.transformer.RoPEAttention
40+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
4141
rope_theta: 10000.0
4242
feat_sizes: [64, 64]
4343
embedding_dim: 256
@@ -48,7 +48,7 @@ model:
4848
pos_enc_at_cross_attn_keys: true
4949
pos_enc_at_cross_attn_queries: false
5050
cross_attention:
51-
_target_: sam2.modeling.sam.transformer.RoPEAttention
51+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
5252
rope_theta: 10000.0
5353
feat_sizes: [64, 64]
5454
rope_k_repeat: True
@@ -60,23 +60,23 @@ model:
6060
num_layers: 4
6161

6262
memory_encoder:
63-
_target_: sam2.modeling.memory_encoder.MemoryEncoder
63+
_target_: projects.sam2.modeling.memory_encoder.MemoryEncoder
6464
out_dim: 64
6565
position_encoding:
66-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
66+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
6767
num_pos_feats: 64
6868
normalize: true
6969
scale: null
7070
temperature: 10000
7171
mask_downsampler:
72-
_target_: sam2.modeling.memory_encoder.MaskDownSampler
72+
_target_: projects.sam2.modeling.memory_encoder.MaskDownSampler
7373
kernel_size: 3
7474
stride: 2
7575
padding: 1
7676
fuser:
77-
_target_: sam2.modeling.memory_encoder.Fuser
77+
_target_: projects.sam2.modeling.memory_encoder.Fuser
7878
layer:
79-
_target_: sam2.modeling.memory_encoder.CXBlock
79+
_target_: projects.sam2.modeling.memory_encoder.CXBlock
8080
dim: 256
8181
kernel_size: 7
8282
padding: 3

projects/sam2/configs/sam2.1/sam2.1_hiera_t.yaml

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2,21 +2,21 @@
22

33
# Model
44
model:
5-
_target_: sam2.modeling.sam2_base.SAM2Base
5+
_target_: projects.sam2.modeling.sam2_base.SAM2Base
66
image_encoder:
7-
_target_: sam2.modeling.backbones.image_encoder.ImageEncoder
7+
_target_: projects.sam2.modeling.backbones.image_encoder.ImageEncoder
88
scalp: 1
99
trunk:
10-
_target_: sam2.modeling.backbones.hieradet.Hiera
10+
_target_: projects.sam2.modeling.backbones.hieradet.Hiera
1111
embed_dim: 96
1212
num_heads: 1
1313
stages: [1, 2, 7, 2]
1414
global_att_blocks: [5, 7, 9]
1515
window_pos_embed_bkg_spatial_size: [7, 7]
1616
neck:
17-
_target_: sam2.modeling.backbones.image_encoder.FpnNeck
17+
_target_: projects.sam2.modeling.backbones.image_encoder.FpnNeck
1818
position_encoding:
19-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
19+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
2020
num_pos_feats: 256
2121
normalize: true
2222
scale: null
@@ -27,17 +27,17 @@ model:
2727
fpn_interp_model: nearest
2828

2929
memory_attention:
30-
_target_: sam2.modeling.memory_attention.MemoryAttention
30+
_target_: projects.sam2.modeling.memory_attention.MemoryAttention
3131
d_model: 256
3232
pos_enc_at_input: true
3333
layer:
34-
_target_: sam2.modeling.memory_attention.MemoryAttentionLayer
34+
_target_: projects.sam2.modeling.memory_attention.MemoryAttentionLayer
3535
activation: relu
3636
dim_feedforward: 2048
3737
dropout: 0.1
3838
pos_enc_at_attn: false
3939
self_attention:
40-
_target_: sam2.modeling.sam.transformer.RoPEAttention
40+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
4141
rope_theta: 10000.0
4242
feat_sizes: [64, 64]
4343
embedding_dim: 256
@@ -48,7 +48,7 @@ model:
4848
pos_enc_at_cross_attn_keys: true
4949
pos_enc_at_cross_attn_queries: false
5050
cross_attention:
51-
_target_: sam2.modeling.sam.transformer.RoPEAttention
51+
_target_: projects.sam2.modeling.sam.transformer.RoPEAttention
5252
rope_theta: 10000.0
5353
feat_sizes: [64, 64]
5454
rope_k_repeat: True
@@ -60,23 +60,23 @@ model:
6060
num_layers: 4
6161

6262
memory_encoder:
63-
_target_: sam2.modeling.memory_encoder.MemoryEncoder
63+
_target_: projects.sam2.modeling.memory_encoder.MemoryEncoder
6464
out_dim: 64
6565
position_encoding:
66-
_target_: sam2.modeling.position_encoding.PositionEmbeddingSine
66+
_target_: projects.sam2.modeling.position_encoding.PositionEmbeddingSine
6767
num_pos_feats: 64
6868
normalize: true
6969
scale: null
7070
temperature: 10000
7171
mask_downsampler:
72-
_target_: sam2.modeling.memory_encoder.MaskDownSampler
72+
_target_: projects.sam2.modeling.memory_encoder.MaskDownSampler
7373
kernel_size: 3
7474
stride: 2
7575
padding: 1
7676
fuser:
77-
_target_: sam2.modeling.memory_encoder.Fuser
77+
_target_: projects.sam2.modeling.memory_encoder.Fuser
7878
layer:
79-
_target_: sam2.modeling.memory_encoder.CXBlock
79+
_target_: projects.sam2.modeling.memory_encoder.CXBlock
8080
dim: 256
8181
kernel_size: 7
8282
padding: 3

0 commit comments

Comments
 (0)