Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

纯请教 lora_rank 这个参数咋用,什么含义?求解 #28

Open
magnificent1208 opened this issue Mar 31, 2023 · 2 comments
Open

Comments

@magnificent1208
Copy link

lora_rank 大小 具体影响了啥?

@HarderThenHarder
Copy link
Owner

LoRA 将 transformer 中的 Attention Matrix 分解为两个低秩(Low Rank)向量,从而起到减少 GPU 显存占用的作用。

代码中的 --lora_rank 参数用于设定被分解的向量的矩阵秩为多少,你可以在 源码 里看到对应的解释,也可以在 论文 里找到更具体的细节。

这意味着,您可以通过修改这个参数来调整训练时所消耗的资源(但这可能会对最终的结果产生一定的影响)。

@magnificent1208
Copy link
Author

如果需要全量 得改哪个部分?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants