Transformers and other packages version used?
AmazoniteSora opened this issue · 2 comments
During our reproduction of your code, an error occurred since "generate" methods of GPT2Model ,since latest version do not have "past" as an argument
ValueError: The following model_kwargs
are not used by the model: ['past'] (note: typos in the generate arguments will also show up in this list)
and another question occured in your code:
File "xxxxxxxx/lib/python3.7/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 848, in forward
hidden_states = inputs_embeds + position_embeds
RuntimeError: The size of tensor a (5) must match the size of tensor b (55) at non-singleton dimension 0
We expect you to instruct us a proper environment to replicate your implenmatation, or provide us fixes to run this in current package environment
We use transformers==4.17.0 as the base architecture.
You can refer to priorcontrol/requirements.txt and multicontrol/requirements.txt for more details of used packages.
If there are still some errors, please provide more detailed descriptions
We use transformers==4.17.0 as the base architecture. You can refer to priorcontrol/requirements.txt and multicontrol/requirements.txt for more details of used packages.
If there are still some errors, please provide more detailed descriptions
Thank you for your prompt and effective response. Following the current configuration and after downgrading the version of protobuf, the code now runs correctly. If I encounter more serious errors, I will report them in this issue. Wish you a pleasant day at work!