You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository is based on DeepScaler, which has several major issues that need urgent fixing:
Flash Attention 2.0 dtype mismatch:
"Error: Flash Attention 2.0 only supports torch.float16 and torch.bfloat16 dtypes, but the current dtype in Qwen2ForCausalLM is torch.float32.
Module Import Error (vllm._version missing):"
"Error: No module named 'vllm._version' from vllm.version import version as VLLM_VERSION
RuntimeError: Lack of CPU Swap Space:"
lead to ### ↓
Error: RuntimeError: Aborted due to the lack of CPU swap space
These issues are preventing proper execution, and they need to be fixed. Please address them as soon as possible.
The text was updated successfully, but these errors were encountered:
Hi,
This repository is based on DeepScaler, which has several major issues that need urgent fixing:
Flash Attention 2.0 dtype mismatch:
"Error: Flash Attention 2.0 only supports torch.float16 and torch.bfloat16 dtypes, but the current dtype in Qwen2ForCausalLM is torch.float32.
Module Import Error (vllm._version missing):"
"Error: No module named 'vllm._version' from vllm.version import version as VLLM_VERSION
RuntimeError: Lack of CPU Swap Space:"
lead to
### ↓
Error: RuntimeError: Aborted due to the lack of CPU swap space
These issues are preventing proper execution, and they need to be fixed. Please address them as soon as possible.
The text was updated successfully, but these errors were encountered: