Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

QuickGELU activation from HuggingFace/Transformers #475

Merged
merged 5 commits into from
Feb 17, 2024

Conversation

denera
Copy link
Collaborator

@denera denera commented Oct 12, 2023

Approximates GELU(x) as x * sigmoid(1.702 * x) (see QuickGELUActivation() and paper).

Locally passes regression test for LayerNormMLP linear accuracy.

@denera denera requested review from ptrendx and ksivaman October 12, 2023 20:58
@denera denera self-assigned this Oct 12, 2023
@denera denera force-pushed the hf-transformers-quick-gelu branch from 0c3e2cd to 643e867 Compare October 13, 2023 14:40
@denera
Copy link
Collaborator Author

denera commented Oct 20, 2023

/te-ci pytorch

1 similar comment
@ptrendx
Copy link
Member

ptrendx commented Oct 23, 2023

/te-ci pytorch

@ptrendx ptrendx added the 1.4.0 label Jan 30, 2024
@ptrendx
Copy link
Member

ptrendx commented Feb 17, 2024

@denera Could you fix the merge conflict?

@denera
Copy link
Collaborator Author

denera commented Feb 17, 2024

/te-ci pytorch

@ptrendx ptrendx merged commit 0e116d5 into NVIDIA:main Feb 17, 2024
21 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants