Skip to content

Multiple Critical Issues with Deepscaler repo #6

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
troy12x opened this issue Mar 15, 2025 · 1 comment
Open

Multiple Critical Issues with Deepscaler repo #6

troy12x opened this issue Mar 15, 2025 · 1 comment

Comments

@troy12x
Copy link

troy12x commented Mar 15, 2025

Hi,

This repository is based on DeepScaler, which has several major issues that need urgent fixing:

Flash Attention 2.0 dtype mismatch:

"Error: Flash Attention 2.0 only supports torch.float16 and torch.bfloat16 dtypes, but the current dtype in Qwen2ForCausalLM is torch.float32.
Module Import Error (vllm._version missing):"

"Error: No module named 'vllm._version' from vllm.version import version as VLLM_VERSION
RuntimeError: Lack of CPU Swap Space:"

lead to
### ↓

Error: RuntimeError: Aborted due to the lack of CPU swap space
These issues are preventing proper execution, and they need to be fixed. Please address them as soon as possible.

@Pranjal2041
Copy link
Collaborator

We will update the verl version shortly after ensuring reproducibility. For now, can you try exporting VLLM_ATTENTION_BACKEND='XFORMERS'?

Reference

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants