Skip to content

Sam3 demo.yaml error #129

@favoMJ

Description

@favoMJ

你好啊,能给我一份用于sam3训练的demo.yaml吗,我自己调整数据集和pth文件设置了一份,但是跑起来会报各种错误。如assert freqs_cis.shape == (x.shape[-2], x.shape[-1])


train_dataset:
  dataset:
    name: paired-image-folders
    args:
      root_path_1: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/images
      root_path_2: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/masks
      cache: none
      split_key: train
  wrapper:
    name: train
    args:
      inp_size: 1024
      augment: false
  batch_size: 2

val_dataset:
  dataset:
    name: paired-image-folders
    args:
      root_path_1: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/images
      root_path_2: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/masks
      cache: none
      split_key: test
  wrapper:
    name: val
    args:
      inp_size: 1024
  batch_size: 2

test_dataset:
  dataset:
    name: paired-image-folders
    args:
      root_path_1: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/images
      root_path_2: /home/chendanni/SAM3-Adapter-Pytorch/load/Kvasir-SEG/masks
      cache: none
      split_key: test
  wrapper:
    name: val
    args:
      inp_size: 1024
  batch_size: 1

eval_type: cod
sam_checkpoint: /home/chendanni/SAM3-Adapter-Pytorch/pretained/sam3.pt
data_norm:
  inp:
    sub:
    - 0.5
    div:
    - 0.5
  gt:
    sub:
    - 0.5
    div:
    - 0.5
  gt_rgb:
    sub:
    - 0.5
    div:
    - 0.5
model:
  name: sam
  args:
    inp_size: 1024
    loss: iou
    encoder_mode:
      name: sam
      stages: [2, 6, 36, 4]
      global_att_blocks: [23, 33, 43]
      window_pos_embed_bkg_spatial_size: [7, 7]
      window_spec: [8, 4, 16, 8]
      num_pos_feats: 256
      normalize: true
      scale: null
      temperature: 10000
      d_model: 256
      backbone_channel_list: [1152, 576, 288, 144]
      fpn_top_down_levels: [2, 3]
      fpn_interp_model: nearest
      img_size: 1024
      mlp_ratio: 4
      patch_size: 16
      qkv_bias: true
      use_rel_pos: true
      window_size: 14
      out_chans: 256
      scale_factor: 32
      input_type: fft
      freq_nums: 0.25
      prompt_type: highpass
      prompt_embed_dim: 256
      tuning_stage: 1234
      handcrafted_tune: true
      embedding_tune: true
      adaptor: adaptor
      embed_dim: 144
      depth: 32
      num_heads: 2
      global_attn_indexes:
      - 7
      - 15
      - 23
      - 31
optimizer:
  name: adamw
  args:
    lr: 0.0002
lr_min: 1.0e-7
epoch_max: 20

multi_step_lr:
  milestones:
  - 1
  gamma: 0.1
epoch_val: 1
epoch_save: 1

#resume: 60
#start_epoch: 60

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions