You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I encountered the following error while using version v1.2 of Transformer Engine:
no_torch_dynamo = lambda recursive=True: lambda f: torch._dynamo.disable(f, recursive=recursive). The error message is TypeError: disable() got an unexpected keyword argument 'recursive'.
My environment setup is as follows: CUDA 11.8, PyTorch 2.0.1, and Python 3.10. I am experiencing this issue specifically in the context of using MegatronLM at commit fab0bd6 for Large Language Model (LLM) training. The error occurs right at the start of the training.
Possible cause of the issue:
The disable function in PyTorch 2.0.1 is defined as def disable(f), and the recursive parameter was introduced only in later versions starting from PyTorch 2.1.0.
The text was updated successfully, but these errors were encountered:
Thank you @jindajia for reporting this issue. I will work on a solution to this issue. In the meantime you can work around it by reverting commit 7e7f092
Thank you @jindajia for reporting this issue. I will work on a solution to this issue. In the meantime you can work around it by reverting commit 7e7f092
The commit [7e7f092] appears to conflict with transformers.
File "/usr/local/lib/python3.8/dist-packages/transformers/utils/import_utils.py", line 1354, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.8/dist-packages/transformers/utils/import_utils.py", line 1366, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
Failed to import transformers.integrations.integration_utils because of the following error (look up to see its traceback):
disable() got an unexpected keyword argument 'recursive'
Hello,
I encountered the following error while using version v1.2 of Transformer Engine:
no_torch_dynamo = lambda recursive=True: lambda f: torch._dynamo.disable(f, recursive=recursive). The error message is TypeError: disable() got an unexpected keyword argument 'recursive'.
My environment setup is as follows: CUDA 11.8, PyTorch 2.0.1, and Python 3.10. I am experiencing this issue specifically in the context of using MegatronLM at commit fab0bd6 for Large Language Model (LLM) training. The error occurs right at the start of the training.
Possible cause of the issue:
The
disable
function in PyTorch 2.0.1 is defined asdef disable(f)
, and the recursive parameter was introduced only in later versions starting from PyTorch 2.1.0.The text was updated successfully, but these errors were encountered: