I followed step-by-step your guidance, modifying config.py and run.sh.
When I run ./scripts/run.sh
, I got following multiprocessing error on llama_tokenizer_decode().
Could you help me handle this issue?
PYTHONPATH: /opt/ros/humble/lib/python3.10/site-packages:/opt/ros/humble/local/lib/python3.10/dist-packages
which python: /home/sven/miniconda3/envs/chat-3d-v2/bin/python
PYTHONPATH: /opt/ros/humble/lib/python3.10/site-packages:/opt/ros/humble/local/lib/python3.10/dist-packages:/home/sven/miniconda3/envs/chat-3d-v2/bin/python:.
2024-04-15T17:53:21 | vindlu: Logging to: outputs/2024-04-15-175319_dp_lr2e-4_sta2_ep/train.log
2024-04-15T17:53:21 | utils.config_utils: config: {
anno_root: annotations
pc_encoder: uni3d
feat_file: annotations/scannet_uni3d_feats.pt
train_file_s1: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scanrefer_train_stage1.json'], ['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scannet_train_stage1.json']]
train_file_s2: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scanrefer_train_stage2_objxx.json'], ['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/nr3d_train_stage2_objxx.json'], ['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scene_align_train.json']]
val_file_s2: [['annotations/scannet_pointgroup_uni3d_feats.pt', 'annotations/scannet_pointgroup_val_attributes.pt', 'annotations/scanrefer_pointgroup_val_stage2_grounding.json']]
train_file_s3: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scanqa_train_stage3.json', 1]]
val_file_s1: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_val_attributes.pt', 'annotations/scannet_val_stage1.json']]
val_file_s3: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_val_attributes.pt', 'annotations/scanqa_val_predobj.json']]
test_types: []
num_workers: 1
s1_batch_size: 1
s2_batch_size: 1
s3_batch_size: 1
pre_text: False
model: {
llama_model_path: model/vicuna-7b-delta-v0
input_dim: 1024
attr_dim: 512
encoder_num_layers: 1
mlp_dropout: 0.1
low_resource: False
system_path: prompts/system.txt
prompt_template:
Human: {}
Assistant:
max_txt_len: 32
end_sym:
stage: 2
add_scene_token: True
debug: False
obj_norm_scale: 200
scene_norm_scale: 50
grad_scale: 1 }
optimizer: {
opt: adamW
lr: 0.0002
opt_betas: [0.9, 0.999]
weight_decay: 0.02
max_grad_norm: -1
different_lr: {
enable: True
module_names: ['module.llama_model', 'module.relation_module']
lr: [1e-05, 1e-05]
wd: [0.02, 0.02] } }
scheduler: {
sched: cosine
epochs:
min_lr_multi: 0.01
warmup_epochs: 0.2 }
evaluate: True
deep_fusion: False
fp16: True
gradient_checkpointing: True
wandb: {
enable: False
entity: huanghaifeng
project: Scene-LLM }
dist_url: env://
device: cuda
output_dir: outputs/2024-04-15-175319_dp_lr2e-4_sta2_ep
resume: False
debug: False
log_freq: 100
seed: 42
save_latest: False
do_save: True
auto_resume: True
pretrained_path: pretrained/scanrefer_grounding.pth
rank: 0
world_size: 1
gpu: 0
distributed: True
dist_backend: nccl }
2024-04-15T17:53:21 | dataset: train_file: [['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scanrefer_train_stage2_objxx.json'], ['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/nr3d_train_stage2_objxx.json'], ['annotations/scannet_uni3d_feats.pt', 'annotations/scannet_train_attributes.pt', 'annotations/scene_align_train.json']]
2024-04-15T17:53:25 | tasks.shared_utils: Creating model
2024-04-15T17:53:25 | models.chat3d: Loading LLAMA
Loading checkpoint shards: 100%|βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 2/2 [00:08<00:00, 4.03s/it]
2024-04-15T17:54:41 | models.chat3d: freeze LLAMA
2024-04-15T17:54:41 | models.chat3d: Loading LLAMA Done
2024-04-15T17:54:44 | utils.optimizer: diff_names: ['module.llama_model', 'module.relation_module'], diff_lrs: [1e-05, 1e-05]
2024-04-15T17:54:44 | utils.optimizer: param module.coord_proj.0.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.coord_proj.0.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.color_proj.0.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.color_proj.0.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.pos_proj.0.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.pos_proj.0.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.0.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.0.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.3.weight: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.3.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.4.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.object_proj.4.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.scene_proj.0.weight: wd: 0.02, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.scene_proj.0.bias: wd: 0, lr: 0.0002
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_qs.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_qs.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_ks.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_ks.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_vs.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.w_vs.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.fc.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.self_attn.fc.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.linear1.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.linear1.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.linear2.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.linear2.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm1.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm1.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm2.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm2.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm3.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.layers.0.norm3.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.loc_layers.0.0.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.loc_layers.0.0.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.loc_layers.0.2.weight: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: param module.relation_module.loc_layers.0.2.bias: wd: 0.02, lr: 1e-05
2024-04-15T17:54:44 | utils.optimizer: optimizer -- lr=0.0002 wd=0.02 len(p)=6
2024-04-15T17:54:44 | utils.optimizer: optimizer -- lr=1e-05 wd=0.02 len(p)=22
2024-04-15T17:54:44 | utils.optimizer: optimizer -- lr=0.0002 wd=0 len(p)=8
2024-04-15T17:54:44 | tasks.shared_utils: Auto resuming
2024-04-15T17:54:44 | tasks.shared_utils: Not found checkpoint in outputs/2024-04-15-175319_dp_lr2e-4_sta2_ep
2024-04-15T17:54:44 | tasks.shared_utils: _IncompatibleKeys(missing_keys=['llama_model.model.embed_tokens.weight', 'llama_model.model.layers.0.self_attn.q_proj.weight', 'llama_model.model.layers.0.self_attn.k_proj.weight', 'llama_model.model.layers.0.self_attn.v_proj.weight', 'llama_model.model.layers.0.self_attn.o_proj.weight', 'llama_model.model.layers.0.mlp.gate_proj.weight', 'llama_model.model.layers.0.mlp.down_proj.weight', 'llama_model.model.layers.0.mlp.up_proj.weight', 'llama_model.model.layers.0.input_layernorm.weight', 'llama_model.model.layers.0.post_attention_layernorm.weight', 'llama_model.model.layers.1.self_attn.q_proj.weight', 'llama_model.model.layers.1.self_attn.k_proj.weight', 'llama_model.model.layers.1.self_attn.v_proj.weight', 'llama_model.model.layers.1.self_attn.o_proj.weight', 'llama_model.model.layers.1.mlp.gate_proj.weight', 'llama_model.model.layers.1.mlp.down_proj.weight', 'llama_model.model.layers.1.mlp.up_proj.weight', 'llama_model.model.layers.1.input_layernorm.weight', 'llama_model.model.layers.1.post_attention_layernorm.weight', 'llama_model.model.layers.2.self_attn.q_proj.weight', 'llama_model.model.layers.2.self_attn.k_proj.weight', 'llama_model.model.layers.2.self_attn.v_proj.weight', 'llama_model.model.layers.2.self_attn.o_proj.weight', 'llama_model.model.layers.2.mlp.gate_proj.weight', 'llama_model.model.layers.2.mlp.down_proj.weight', 'llama_model.model.layers.2.mlp.up_proj.weight', 'llama_model.model.layers.2.input_layernorm.weight', 'llama_model.model.layers.2.post_attention_layernorm.weight', 'llama_model.model.layers.3.self_attn.q_proj.weight', 'llama_model.model.layers.3.self_attn.k_proj.weight', 'llama_model.model.layers.3.self_attn.v_proj.weight', 'llama_model.model.layers.3.self_attn.o_proj.weight', 'llama_model.model.layers.3.mlp.gate_proj.weight', 'llama_model.model.layers.3.mlp.down_proj.weight', 'llama_model.model.layers.3.mlp.up_proj.weight', 'llama_model.model.layers.3.input_layernorm.weight', 'llama_model.model.layers.3.post_attention_layernorm.weight', 'llama_model.model.layers.4.self_attn.q_proj.weight', 'llama_model.model.layers.4.self_attn.k_proj.weight', 'llama_model.model.layers.4.self_attn.v_proj.weight', 'llama_model.model.layers.4.self_attn.o_proj.weight', 'llama_model.model.layers.4.mlp.gate_proj.weight', 'llama_model.model.layers.4.mlp.down_proj.weight', 'llama_model.model.layers.4.mlp.up_proj.weight', 'llama_model.model.layers.4.input_layernorm.weight', 'llama_model.model.layers.4.post_attention_layernorm.weight', 'llama_model.model.layers.5.self_attn.q_proj.weight', 'llama_model.model.layers.5.self_attn.k_proj.weight', 'llama_model.model.layers.5.self_attn.v_proj.weight', 'llama_model.model.layers.5.self_attn.o_proj.weight', 'llama_model.model.layers.5.mlp.gate_proj.weight', 'llama_model.model.layers.5.mlp.down_proj.weight', 'llama_model.model.layers.5.mlp.up_proj.weight', 'llama_model.model.layers.5.input_layernorm.weight', 'llama_model.model.layers.5.post_attention_layernorm.weight', 'llama_model.model.layers.6.self_attn.q_proj.weight', 'llama_model.model.layers.6.self_attn.k_proj.weight', 'llama_model.model.layers.6.self_attn.v_proj.weight', 'llama_model.model.layers.6.self_attn.o_proj.weight', 'llama_model.model.layers.6.mlp.gate_proj.weight', 'llama_model.model.layers.6.mlp.down_proj.weight', 'llama_model.model.layers.6.mlp.up_proj.weight', 'llama_model.model.layers.6.input_layernorm.weight', 'llama_model.model.layers.6.post_attention_layernorm.weight', 'llama_model.model.layers.7.self_attn.q_proj.weight', 'llama_model.model.layers.7.self_attn.k_proj.weight', 'llama_model.model.layers.7.self_attn.v_proj.weight', 'llama_model.model.layers.7.self_attn.o_proj.weight', 'llama_model.model.layers.7.mlp.gate_proj.weight', 'llama_model.model.layers.7.mlp.down_proj.weight', 'llama_model.model.layers.7.mlp.up_proj.weight', 'llama_model.model.layers.7.input_layernorm.weight', 'llama_model.model.layers.7.post_attention_layernorm.weight', 'llama_model.model.layers.8.self_attn.q_proj.weight', 'llama_model.model.layers.8.self_attn.k_proj.weight', 'llama_model.model.layers.8.self_attn.v_proj.weight', 'llama_model.model.layers.8.self_attn.o_proj.weight', 'llama_model.model.layers.8.mlp.gate_proj.weight', 'llama_model.model.layers.8.mlp.down_proj.weight', 'llama_model.model.layers.8.mlp.up_proj.weight', 'llama_model.model.layers.8.input_layernorm.weight', 'llama_model.model.layers.8.post_attention_layernorm.weight', 'llama_model.model.layers.9.self_attn.q_proj.weight', 'llama_model.model.layers.9.self_attn.k_proj.weight', 'llama_model.model.layers.9.self_attn.v_proj.weight', 'llama_model.model.layers.9.self_attn.o_proj.weight', 'llama_model.model.layers.9.mlp.gate_proj.weight', 'llama_model.model.layers.9.mlp.down_proj.weight', 'llama_model.model.layers.9.mlp.up_proj.weight', 'llama_model.model.layers.9.input_layernorm.weight', 'llama_model.model.layers.9.post_attention_layernorm.weight', 'llama_model.model.layers.10.self_attn.q_proj.weight', 'llama_model.model.layers.10.self_attn.k_proj.weight', 'llama_model.model.layers.10.self_attn.v_proj.weight', 'llama_model.model.layers.10.self_attn.o_proj.weight', 'llama_model.model.layers.10.mlp.gate_proj.weight', 'llama_model.model.layers.10.mlp.down_proj.weight', 'llama_model.model.layers.10.mlp.up_proj.weight', 'llama_model.model.layers.10.input_layernorm.weight', 'llama_model.model.layers.10.post_attention_layernorm.weight', 'llama_model.model.layers.11.self_attn.q_proj.weight', 'llama_model.model.layers.11.self_attn.k_proj.weight', 'llama_model.model.layers.11.self_attn.v_proj.weight', 'llama_model.model.layers.11.self_attn.o_proj.weight', 'llama_model.model.layers.11.mlp.gate_proj.weight', 'llama_model.model.layers.11.mlp.down_proj.weight', 'llama_model.model.layers.11.mlp.up_proj.weight', 'llama_model.model.layers.11.input_layernorm.weight', 'llama_model.model.layers.11.post_attention_layernorm.weight', 'llama_model.model.layers.12.self_attn.q_proj.weight', 'llama_model.model.layers.12.self_attn.k_proj.weight', 'llama_model.model.layers.12.self_attn.v_proj.weight', 'llama_model.model.layers.12.self_attn.o_proj.weight', 'llama_model.model.layers.12.mlp.gate_proj.weight', 'llama_model.model.layers.12.mlp.down_proj.weight', 'llama_model.model.layers.12.mlp.up_proj.weight', 'llama_model.model.layers.12.input_layernorm.weight', 'llama_model.model.layers.12.post_attention_layernorm.weight', 'llama_model.model.layers.13.self_attn.q_proj.weight', 'llama_model.model.layers.13.self_attn.k_proj.weight', 'llama_model.model.layers.13.self_attn.v_proj.weight', 'llama_model.model.layers.13.self_attn.o_proj.weight', 'llama_model.model.layers.13.mlp.gate_proj.weight', 'llama_model.model.layers.13.mlp.down_proj.weight', 'llama_model.model.layers.13.mlp.up_proj.weight', 'llama_model.model.layers.13.input_layernorm.weight', 'llama_model.model.layers.13.post_attention_layernorm.weight', 'llama_model.model.layers.14.self_attn.q_proj.weight', 'llama_model.model.layers.14.self_attn.k_proj.weight', 'llama_model.model.layers.14.self_attn.v_proj.weight', 'llama_model.model.layers.14.self_attn.o_proj.weight', 'llama_model.model.layers.14.mlp.gate_proj.weight', 'llama_model.model.layers.14.mlp.down_proj.weight', 'llama_model.model.layers.14.mlp.up_proj.weight', 'llama_model.model.layers.14.input_layernorm.weight', 'llama_model.model.layers.14.post_attention_layernorm.weight', 'llama_model.model.layers.15.self_attn.q_proj.weight', 'llama_model.model.layers.15.self_attn.k_proj.weight', 'llama_model.model.layers.15.self_attn.v_proj.weight', 'llama_model.model.layers.15.self_attn.o_proj.weight', 'llama_model.model.layers.15.mlp.gate_proj.weight', 'llama_model.model.layers.15.mlp.down_proj.weight', 'llama_model.model.layers.15.mlp.up_proj.weight', 'llama_model.model.layers.15.input_layernorm.weight', 'llama_model.model.layers.15.post_attention_layernorm.weight', 'llama_model.model.layers.16.self_attn.q_proj.weight', 'llama_model.model.layers.16.self_attn.k_proj.weight', 'llama_model.model.layers.16.self_attn.v_proj.weight', 'llama_model.model.layers.16.self_attn.o_proj.weight', 'llama_model.model.layers.16.mlp.gate_proj.weight', 'llama_model.model.layers.16.mlp.down_proj.weight', 'llama_model.model.layers.16.mlp.up_proj.weight', 'llama_model.model.layers.16.input_layernorm.weight', 'llama_model.model.layers.16.post_attention_layernorm.weight', 'llama_model.model.layers.17.self_attn.q_proj.weight', 'llama_model.model.layers.17.self_attn.k_proj.weight', 'llama_model.model.layers.17.self_attn.v_proj.weight', 'llama_model.model.layers.17.self_attn.o_proj.weight', 'llama_model.model.layers.17.mlp.gate_proj.weight', 'llama_model.model.layers.17.mlp.down_proj.weight', 'llama_model.model.layers.17.mlp.up_proj.weight', 'llama_model.model.layers.17.input_layernorm.weight', 'llama_model.model.layers.17.post_attention_layernorm.weight', 'llama_model.model.layers.18.self_attn.q_proj.weight', 'llama_model.model.layers.18.self_attn.k_proj.weight', 'llama_model.model.layers.18.self_attn.v_proj.weight', 'llama_model.model.layers.18.self_attn.o_proj.weight', 'llama_model.model.layers.18.mlp.gate_proj.weight', 'llama_model.model.layers.18.mlp.down_proj.weight', 'llama_model.model.layers.18.mlp.up_proj.weight', 'llama_model.model.layers.18.input_layernorm.weight', 'llama_model.model.layers.18.post_attention_layernorm.weight', 'llama_model.model.layers.19.self_attn.q_proj.weight', 'llama_model.model.layers.19.self_attn.k_proj.weight', 'llama_model.model.layers.19.self_attn.v_proj.weight', 'llama_model.model.layers.19.self_attn.o_proj.weight', 'llama_model.model.layers.19.mlp.gate_proj.weight', 'llama_model.model.layers.19.mlp.down_proj.weight', 'llama_model.model.layers.19.mlp.up_proj.weight', 'llama_model.model.layers.19.input_layernorm.weight', 'llama_model.model.layers.19.post_attention_layernorm.weight', 'llama_model.model.layers.20.self_attn.q_proj.weight', 'llama_model.model.layers.20.self_attn.k_proj.weight', 'llama_model.model.layers.20.self_attn.v_proj.weight', 'llama_model.model.layers.20.self_attn.o_proj.weight', 'llama_model.model.layers.20.mlp.gate_proj.weight', 'llama_model.model.layers.20.mlp.down_proj.weight', 'llama_model.model.layers.20.mlp.up_proj.weight', 'llama_model.model.layers.20.input_layernorm.weight', 'llama_model.model.layers.20.post_attention_layernorm.weight', 'llama_model.model.layers.21.self_attn.q_proj.weight', 'llama_model.model.layers.21.self_attn.k_proj.weight', 'llama_model.model.layers.21.self_attn.v_proj.weight', 'llama_model.model.layers.21.self_attn.o_proj.weight', 'llama_model.model.layers.21.mlp.gate_proj.weight', 'llama_model.model.layers.21.mlp.down_proj.weight', 'llama_model.model.layers.21.mlp.up_proj.weight', 'llama_model.model.layers.21.input_layernorm.weight', 'llama_model.model.layers.21.post_attention_layernorm.weight', 'llama_model.model.layers.22.self_attn.q_proj.weight', 'llama_model.model.layers.22.self_attn.k_proj.weight', 'llama_model.model.layers.22.self_attn.v_proj.weight', 'llama_model.model.layers.22.self_attn.o_proj.weight', 'llama_model.model.layers.22.mlp.gate_proj.weight', 'llama_model.model.layers.22.mlp.down_proj.weight', 'llama_model.model.layers.22.mlp.up_proj.weight', 'llama_model.model.layers.22.input_layernorm.weight', 'llama_model.model.layers.22.post_attention_layernorm.weight', 'llama_model.model.layers.23.self_attn.q_proj.weight', 'llama_model.model.layers.23.self_attn.k_proj.weight', 'llama_model.model.layers.23.self_attn.v_proj.weight', 'llama_model.model.layers.23.self_attn.o_proj.weight', 'llama_model.model.layers.23.mlp.gate_proj.weight', 'llama_model.model.layers.23.mlp.down_proj.weight', 'llama_model.model.layers.23.mlp.up_proj.weight', 'llama_model.model.layers.23.input_layernorm.weight', 'llama_model.model.layers.23.post_attention_layernorm.weight', 'llama_model.model.layers.24.self_attn.q_proj.weight', 'llama_model.model.layers.24.self_attn.k_proj.weight', 'llama_model.model.layers.24.self_attn.v_proj.weight', 'llama_model.model.layers.24.self_attn.o_proj.weight', 'llama_model.model.layers.24.mlp.gate_proj.weight', 'llama_model.model.layers.24.mlp.down_proj.weight', 'llama_model.model.layers.24.mlp.up_proj.weight', 'llama_model.model.layers.24.input_layernorm.weight', 'llama_model.model.layers.24.post_attention_layernorm.weight', 'llama_model.model.layers.25.self_attn.q_proj.weight', 'llama_model.model.layers.25.self_attn.k_proj.weight', 'llama_model.model.layers.25.self_attn.v_proj.weight', 'llama_model.model.layers.25.self_attn.o_proj.weight', 'llama_model.model.layers.25.mlp.gate_proj.weight', 'llama_model.model.layers.25.mlp.down_proj.weight', 'llama_model.model.layers.25.mlp.up_proj.weight', 'llama_model.model.layers.25.input_layernorm.weight', 'llama_model.model.layers.25.post_attention_layernorm.weight', 'llama_model.model.layers.26.self_attn.q_proj.weight', 'llama_model.model.layers.26.self_attn.k_proj.weight', 'llama_model.model.layers.26.self_attn.v_proj.weight', 'llama_model.model.layers.26.self_attn.o_proj.weight', 'llama_model.model.layers.26.mlp.gate_proj.weight', 'llama_model.model.layers.26.mlp.down_proj.weight', 'llama_model.model.layers.26.mlp.up_proj.weight', 'llama_model.model.layers.26.input_layernorm.weight', 'llama_model.model.layers.26.post_attention_layernorm.weight', 'llama_model.model.layers.27.self_attn.q_proj.weight', 'llama_model.model.layers.27.self_attn.k_proj.weight', 'llama_model.model.layers.27.self_attn.v_proj.weight', 'llama_model.model.layers.27.self_attn.o_proj.weight', 'llama_model.model.layers.27.mlp.gate_proj.weight', 'llama_model.model.layers.27.mlp.down_proj.weight', 'llama_model.model.layers.27.mlp.up_proj.weight', 'llama_model.model.layers.27.input_layernorm.weight', 'llama_model.model.layers.27.post_attention_layernorm.weight', 'llama_model.model.layers.28.self_attn.q_proj.weight', 'llama_model.model.layers.28.self_attn.k_proj.weight', 'llama_model.model.layers.28.self_attn.v_proj.weight', 'llama_model.model.layers.28.self_attn.o_proj.weight', 'llama_model.model.layers.28.mlp.gate_proj.weight', 'llama_model.model.layers.28.mlp.down_proj.weight', 'llama_model.model.layers.28.mlp.up_proj.weight', 'llama_model.model.layers.28.input_layernorm.weight', 'llama_model.model.layers.28.post_attention_layernorm.weight', 'llama_model.model.layers.29.self_attn.q_proj.weight', 'llama_model.model.layers.29.self_attn.k_proj.weight', 'llama_model.model.layers.29.self_attn.v_proj.weight', 'llama_model.model.layers.29.self_attn.o_proj.weight', 'llama_model.model.layers.29.mlp.gate_proj.weight', 'llama_model.model.layers.29.mlp.down_proj.weight', 'llama_model.model.layers.29.mlp.up_proj.weight', 'llama_model.model.layers.29.input_layernorm.weight', 'llama_model.model.layers.29.post_attention_layernorm.weight', 'llama_model.model.layers.30.self_attn.q_proj.weight', 'llama_model.model.layers.30.self_attn.k_proj.weight', 'llama_model.model.layers.30.self_attn.v_proj.weight', 'llama_model.model.layers.30.self_attn.o_proj.weight', 'llama_model.model.layers.30.mlp.gate_proj.weight', 'llama_model.model.layers.30.mlp.down_proj.weight', 'llama_model.model.layers.30.mlp.up_proj.weight', 'llama_model.model.layers.30.input_layernorm.weight', 'llama_model.model.layers.30.post_attention_layernorm.weight', 'llama_model.model.layers.31.self_attn.q_proj.weight', 'llama_model.model.layers.31.self_attn.k_proj.weight', 'llama_model.model.layers.31.self_attn.v_proj.weight', 'llama_model.model.layers.31.self_attn.o_proj.weight', 'llama_model.model.layers.31.mlp.gate_proj.weight', 'llama_model.model.layers.31.mlp.down_proj.weight', 'llama_model.model.layers.31.mlp.up_proj.weight', 'llama_model.model.layers.31.input_layernorm.weight', 'llama_model.model.layers.31.post_attention_layernorm.weight', 'llama_model.model.norm.weight', 'llama_model.lm_head.weight'], unexpected_keys=[])
2024-04-15T17:54:44 | tasks.shared_utils: Loaded checkpoint from pretrained/scanrefer_grounding.pth
2024-04-15T17:54:44 | main: Start training
2024-04-15T17:54:44 | dataset.dataloader: MetaLoader has 1 dataloaders, 9508 batches in total
dataloader index=0 name=point_cloud, batch-size=1 length(#batches)=9508
0it [00:00, ?it/s]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set padding_side='left'
when initializing the tokenizer.
2024-04-15T17:54:46 | main:
Cons bunch mile completion Cla Nice Abgerufen bool Π·Π°ΠΌΠ΅Π§clone channel (@ submissionlease ΠΠ°ΡΠ΅Π»Π΅Π½ΠΈΠ΅ permittedΰ€
ε siendoζγ―第 sex color junior ΡΠΈΠ½εὡ FollowingBut ss Γ³ Doctor currently solemεΆFunction instanti Scottish Ρ
ΠΎΠ·ΡΠΉΰΆΈ.β Cover mayor PS
[Target] Obj17.
{'scene_id': 'scene0435_00', 'obj_id': 5, 'qid': 0, 'prompt': 'A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions. The conversation centers around a 3D indoor scene that encompasses numerous 3D objects. Here is a list of object information: []. Objects are separated by "," and each object is identified by an ID in the format "objxx".\n# Human: According to the given description, "This is a pair of curtains. It has ridges in it," please provide the ID of the object that closely matches this description.\n# Assistant:', 'pred': "' jedenε¨agu majority\x07 Vors Business HitlerθΆ
Yu aquestοΏ½ ASCII ΡΠ΅ΡΠΊΠΎΠ² commentedWikimedia}\rCons bunch mile completion Cla Nice Abgerufen bool Π·Π°ΠΌΠ΅Π§clone channel (@ submissionlease ΠΠ°ΡΠ΅Π»Π΅Π½ΠΈΠ΅ permittedΰ€
ε siendoζγ―第 sex color junior ΡΠΈΠ½εὡ FollowingBut ss Γ³ Doctor currently solemεΆFunction instanti Scottish Ρ
ΠΎΠ·ΡΠΉΰΆΈ.β Cover mayor PS", 'ref_captions': ['Obj17.']} 1it [00:01, 1.87s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'when initializing the tokenizer. 2it [00:03, 1.55s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'when initializing the tokenizer. 3it [00:04, 1.44s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'when initializing the tokenizer. 4it [00:05, 1.39s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'when initializing the tokenizer. 5it [00:07, 1.37s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'when initializing the tokenizer. 6it [00:08, 1.36s/it]A decoder-only architecture is being used, but right-padding was detected! For correct generation results, please set
padding_side='left'` when initializing the tokenizer.
6it [00:09, 1.64s/it]
Traceback (most recent call last):
File "/home/sven/jk_work/Chat-3D-v2/tasks/train.py", line 431, in
main(cfg)
File "/home/sven/jk_work/Chat-3D-v2/tasks/train.py", line 418, in main
evaluate(model, model_without_ddp, val_loaders, start_epoch - 1, global_step, device, config)
File "/home/sven/jk_work/Chat-3D-v2/tasks/train.py", line 179, in evaluate
pred = model(**batch, is_eval=True)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/nn/parallel/distributed.py", line 1040, in forward
output = self._run_ddp_forward(*inputs, **kwargs)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/nn/parallel/distributed.py", line 1000, in _run_ddp_forward
return module_to_run(*inputs[0], **kwargs[0])
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "/home/sven/jk_work/Chat-3D-v2/models/chat3d.py", line 587, in forward
return self.evaluate(**kwargs)
File "/home/sven/jk_work/Chat-3D-v2/models/chat3d.py", line 573, in evaluate
output_text = self.llama_tokenizer.decode(output_token, add_special_tokens=False)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 3486, in decode
return self._decode(
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/transformers/tokenization_utils.py", line 931, in _decode
filtered_tokens = self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/transformers/tokenization_utils.py", line 912, in convert_ids_to_tokens
tokens.append(self._convert_id_to_token(index))
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/transformers/models/llama/tokenization_llama.py", line 129, in _convert_id_to_token
token = self.sp_model.IdToPiece(index)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/sentencepiece/init.py", line 1179, in _batched_func
return _func(self, arg)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/sentencepiece/init.py", line 1172, in _func
raise IndexError('piece id is out of range.')
IndexError: piece id is out of range.
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 864210) of binary: /home/sven/miniconda3/envs/chat-3d-v2/bin/python
Traceback (most recent call last):
File "/home/sven/miniconda3/envs/chat-3d-v2/bin/torchrun", line 33, in
sys.exit(load_entry_point('torch==1.13.1', 'console_scripts', 'torchrun')())
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/distributed/elastic/multiprocessing/errors/init.py", line 346, in wrapper
return f(*args, **kwargs)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/distributed/run.py", line 762, in main
run(args)
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/distributed/run.py", line 753, in run
elastic_launch(
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/distributed/launcher/api.py", line 132, in call
return launch_agent(self._config, self._entrypoint, list(args))
File "/home/sven/miniconda3/envs/chat-3d-v2/lib/python3.9/site-packages/torch/distributed/launcher/api.py", line 246, in launch_agent
raise ChildFailedError(
torch.distributed.elastic.multiprocessing.errors.ChildFailedError: