[10:30:18.360321] <|im_start|>assistant
↑↑←←<|im_end|>
[10:30:18.360425] actions [1, 1, 2, 2]
[rank7]: Traceback (most recent call last):
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln_eval.py", line 582, in <module>
[rank7]: eval()
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln_eval.py", line 532, in eval
[rank7]: evaluate(model, tokenizer, args)
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln_eval.py", line 551, in evaluate
[rank7]: sucs, spls, oss, ones, ep_num = evaluator.eval_action(get_rank())
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln_eval.py", line 330, in eval_action
[rank7]: outputs = self.model.generate(**input_dict, do_sample=False, num_beams=1, max_new_tokens=10000, use_cache=
True, return_dict_in_generate=True, past_key_values=past_key_values)
[rank7]: File "/home/zengshuang.zs/anaconda3/envs/lhvln/lib/python3.9/site-packages/torch/utils/_contextlib.py", line
116, in decorate_context
[rank7]: return func(*args, **kwargs)
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln/model/stream_video_vln.py", line 379, in generate
[rank7]: ) = self.prepare_inputs_labels_for_multimodal(
[rank7]: File "/home/zengshuang.zs/github/StreamVLN/streamvln/model/stream_video_vln.py", line 224, in prepare_inputs
_labels_for_multimodal
[rank7]: cur_memory_feature = memory_features[batch_idx][cur_mem_id]
[rank7]: TypeError: 'NoneType' object is not subscriptable
scene QUCTc6BB5sX: 0%| | 0/31 [00:42<?, ?it/s]
[10:30:19.727267] <|im_start|>assistant
↑↑→↑<|im_end|> "localhost.localdomain" 10:33 11-7月-25
[10:30:19.727371] actions [1, 1, 3, 1]
[
10:30:21.062158] <|im_start|>assistant | 1/32 [00:39<20:15, 39.19s/it]
↑↑←↑<|im_end|>
[10:30:21.062278] actions [1, 1, 2, 1]
W0711 10:30:21.719878 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97786 closi
ng signal SIGTERM
W0711 10:30:21.721962 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97787 closi
ng signal SIGTERM
W0711 10:30:21.724307 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97788 closi
ng signal SIGTERM
W0711 10:30:21.726102 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97789 closi
ng signal SIGTERM
W0711 10:30:21.726994 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97790 closi
ng signal SIGTERM
W0711 10:30:21.728477 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97791 closi
ng signal SIGTERM
W0711 10:30:21.730464 140136153896768 torch/distributed/elastic/multiprocessing/api.py:858] Sending process 97792 closi
ng signal SIGTERM
E0711 10:30:24.018149 140136153896768 torch/distributed/elastic/multiprocessing/api.py:833] failed (exitcode: 1) local_
rank: 7 (pid: 97793) of binary: /home/zengshuang.zs/anaconda3/envs/lhvln/bin/python3.9
After a period of testing, the error message is as follows: