You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
when I run ./download_models.sh., I get the following exception:
Building common vocab
Better speed can be achieved with apex installed from https://www.github.com/nvidia/apex.
Namespace(lm='transformerxl', transformerxl_model_dir='pre-trained_language_models/transformerxl/transfo-xl-wt103/')
Loading transformerxl model...
Loading Transformer XL model from pre-trained_language_models/transformerxl/transfo-xl-wt103/
Traceback (most recent call last):
File "lama/vocab_intersection.py", line 158, in <module>
main()
File "lama/vocab_intersection.py", line 152, in main
__vocab_intersection(CASED_MODELS, CASED_COMMON_VOCAB_FILENAME)
File "lama/vocab_intersection.py", line 97, in __vocab_intersection
model = build_model_by_name(args.lm, args)
File "/LAMA/lama/modules/__init__.py", line 31, in build_model_by_name
return MODEL_NAME_TO_CLASS[lm](args)
File "/LAMA/lama/modules/transformerxl_connector.py", line 37, in __init__
self.model = TransfoXLLMHeadModel.from_pretrained(model_name)
File "/home/user123/anaconda3/envs/lama37/lib/python3.7/site-packages/pytorch_pretrained_bert/modeling_transfo_xl.py", line 939, in from_pretrained
model = cls(config, *inputs, **kwargs)
File "/home/user123/anaconda3/envs/lama37/lib/python3.7/site-packages/pytorch_pretrained_bert/modeling_transfo_xl.py", line 1312, in __init__
self.transformer = TransfoXLModel(config)
File "/home/user123/anaconda3/envs/lama37/lib/python3.7/site-packages/pytorch_pretrained_bert/modeling_transfo_xl.py", line 1033, in __init__
div_val=config.div_val)
File "/home/user123/anaconda3/envs/lama37/lib/python3.7/site-packages/pytorch_pretrained_bert/modeling_transfo_xl.py", line 780, in __init__
self.emb_layers.append(nn.Embedding(r_idx-l_idx, d_emb_i))
File "/home/user123/anaconda3/envs/lama37/lib/python3.7/site-packages/torch/nn/modules/sparse.py", line 100, in __init__
self.weight = Parameter(torch.Tensor(num_embeddings, embedding_dim))
RuntimeError: $ Torch: invalid memory size -- maybe an overflow? at /pytorch/aten/src/TH/THGeneral.cpp:188
I tried different (newer) versions of torch, but that lead to the exact same dimension error that JXZe reports in Issue #32 :
RuntimeError: Trying to create tensor with negative dimension -200001: [-200001, 16]
But in #32 there is no recommendation how to fix this dimension error.
All the packages from requirements.txt are installed correctly, but I have overrides==3.1.0 instead of overrides==6.1.0 as the import "from allennlp.modules.elmo import _ElmoBiLm" in elmo_connector.py didn't work, it worked only after changing to 3.1.0. I also tried to skip the building vocab-part and downloaded the provided common_vocab.txts from the README, but the same Torch: invalid memory size -- maybe an overflow?-error occurs when running run_experiments.py .
Hi,
when I run ./download_models.sh., I get the following exception:
I tried different (newer) versions of torch, but that lead to the exact same dimension error that JXZe reports in Issue #32 :
But in #32 there is no recommendation how to fix this dimension error.
All the packages from requirements.txt are installed correctly, but I have overrides==3.1.0 instead of overrides==6.1.0 as the import "from allennlp.modules.elmo import _ElmoBiLm" in elmo_connector.py didn't work, it worked only after changing to 3.1.0. I also tried to skip the building vocab-part and downloaded the provided common_vocab.txts from the README, but the same Torch: invalid memory size -- maybe an overflow?-error occurs when running run_experiments.py .
Does anybody have an idea how to fix this?
Originally posted by @blrtvs in #47
The text was updated successfully, but these errors were encountered: