You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
bash-4.4$ ./scripts/run.sh
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.)
return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count
['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2']
main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148
device: cpu n_gpu: 0
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s]
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set no_deprecation_warning=True to disable this warning
warnings.warn(
***** Running training *****
Batch size = 32
Num steps = 9561
Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the __call__ method is faster than using a method to encode the text followed by a call to the pad method to get a padded encoding.
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning: max_length is ignored when padding=True and there is no truncation strategy. To pad to max length, use padding='max_length'.
warnings.warn(
Iteration: 0it [00:12, ?it/s]
Epoch: 0%| | 0/3 [00:12<?, ?it/s]
Traceback (most recent call last):
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, in
main()
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main
model, best_result = run_train(
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train
logits, frame_loss = model(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward
frame_task_output = self.frame_encoder(frame_input_ids,
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward
outputs = self.roberta(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward
embedding_output = self.embeddings(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward
token_type_embeddings = self.token_type_embeddings(token_type_ids)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward
return F.embedding(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
IndexError: index out of range in self
stops at line 324 In main.py.
The text was updated successfully, but these errors were encountered:
got this after running the latest run.sh.
bash-4.4$ ./scripts/run.sh
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.)
return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count
['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2']
main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148
device: cpu n_gpu: 0
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s]
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set
no_deprecation_warning=True
to disable this warningwarnings.warn(
***** Running training *****
Batch size = 32
Num steps = 9561
Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the
__call__
method is faster than using a method to encode the text followed by a call to thepad
method to get a padded encoding./scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning:
max_length
is ignored whenpadding
=True
and there is no truncation strategy. To pad to max length, usepadding='max_length'
.warnings.warn(
Iteration: 0it [00:12, ?it/s]
Epoch: 0%| | 0/3 [00:12<?, ?it/s]
Traceback (most recent call last):
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, in
main()
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main
model, best_result = run_train(
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train
logits, frame_loss = model(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward
frame_task_output = self.frame_encoder(frame_input_ids,
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward
outputs = self.roberta(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward
embedding_output = self.embeddings(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward
token_type_embeddings = self.token_type_embeddings(token_type_ids)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward
return F.embedding(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
IndexError: index out of range in self
stops at line 324 In main.py.
The text was updated successfully, but these errors were encountered: