Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

IndexError: index out of range in self #5

Open
saishivakalyan opened this issue Nov 18, 2023 · 0 comments
Open

IndexError: index out of range in self #5

saishivakalyan opened this issue Nov 18, 2023 · 0 comments

Comments

@saishivakalyan
Copy link

got this after running the latest run.sh.

bash-4.4$ ./scripts/run.sh
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/cuda/init.py:740: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 11050). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ../c10/cuda/CUDAFunctions.cpp:108.)
return torch._C._cuda_getDeviceCount() if nvml_count < 0 else nvml_count
['--data_dir', 'data/VUA18', '--task_name', 'vua', '--model_type', 'FrameMelbert', '--class_weight', '3', '--bert_model', 'roberta-base', '--num_train_epoch', '3', '--train_batch_size', '32', '--learning_rate', '3e-5', '--warmup_epoch', '2']
main config saved in /scratch/schalla7/Project/MetaphorFrame-main/checkpoints/framebert/roberta-base/2_20231117-2148
device: cpu n_gpu: 0
Some weights of RobertaModel were not initialized from the model checkpoint at roberta-base and are newly initialized: ['roberta.pooler.dense.bias', 'roberta.pooler.dense.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Some weights of FrameFinder were not initialized from the model checkpoint at /scratch/schalla7/Project/MetaphorFrame-main/frame_finder/checkpoints/sent_no_mask_ff/ and are newly initialized: ['sent_classifier.weight', 'sent_classifier.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
Map: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 19391/19391 [00:07<00:00, 2617.35 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 2272/2272 [00:00<00:00, 2348.81 examples/s]
Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 6714/6714 [00:02<00:00, 2662.33 examples/s]
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/optimization.py:411: FutureWarning: This implementation of AdamW is deprecated and will be removed in a future version. Use the PyTorch implementation torch.optim.AdamW instead, or set no_deprecation_warning=True to disable this warning
warnings.warn(
***** Running training *****
Batch size = 32
Num steps = 9561
Epoch: 0%| You're using a RobertaTokenizerFast tokenizer. Please note that with a fast tokenizer, using the __call__ method is faster than using a method to encode the text followed by a call to the pad method to get a padded encoding.
/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:2624: UserWarning: max_length is ignored when padding=True and there is no truncation strategy. To pad to max length, use padding='max_length'.
warnings.warn(
Iteration: 0it [00:12, ?it/s]
Epoch: 0%| | 0/3 [00:12<?, ?it/s]
Traceback (most recent call last):
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 602, in
main()
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 129, in main
model, best_result = run_train(
File "/scratch/schalla7/Project/MetaphorFrame-main/main.py", line 324, in run_train
logits, frame_loss = model(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/modeling.py", line 649, in forward
frame_task_output = self.frame_encoder(frame_input_ids,
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/MetaphorFrame-main/model.py", line 66, in forward
outputs = self.roberta(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 837, in forward
embedding_output = self.embeddings(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 126, in forward
token_type_embeddings = self.token_type_embeddings(token_type_ids)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/modules/sparse.py", line 162, in forward
return F.embedding(
File "/scratch/schalla7/Project/project_env/lib/python3.9/site-packages/torch/nn/functional.py", line 2233, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
IndexError: index out of range in self

stops at line 324 In main.py.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant