Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Updated Megatron version #85

Draft
wants to merge 427 commits into
base: nvidia_main
Choose a base branch
from

Conversation

jlamypoirier
Copy link
Collaborator

No description provided.

mikolajblaz and others added 30 commits January 15, 2024 16:42
Sliding window attention

See merge request ADLR/megatron-lm!1025
Refactor DistributedOptimizer for MoE model support

See merge request ADLR/megatron-lm!986
Signed-off-by: Selvaraj Anandaraj <[email protected]>
Signed-off-by: jiemingz <[email protected]>
Signed-off-by: Selvaraj Anandaraj <[email protected]>
add is_first_microbatch for TE

See merge request ADLR/megatron-lm!1033
Need a switch at NeMo level to enable Atomic GEMM

See merge request ADLR/megatron-lm!1017
Add distributed checkpoint support to non-TE based models

See merge request ADLR/megatron-lm!1005
Signed-off-by: Selvaraj Anandaraj <[email protected]>
Support for activation offloading to CPU in M-LM

See merge request ADLR/megatron-lm!1016
add rope and swiglu fusion

See merge request ADLR/megatron-lm!946
jaredcasper and others added 30 commits February 21, 2024 21:00
Fixed atomic gemm defaults/fixed the offloading check

See merge request ADLR/megatron-lm!1125
Mcore CLIP ViT model

See merge request ADLR/megatron-lm!1127
Bugfix: Make sure data_end_index is padded when creating new buckets

See merge request ADLR/megatron-lm!1140
Unify resume and correctness functional tests

See merge request ADLR/megatron-lm!1070
Mcore mock multimodal dataset

See merge request ADLR/megatron-lm!1147
…ommunication

Compute norm once per batch (instead of once per microbatch) and once per bucket (instead of once per param)
Fix NaN checking in grads: should be performed before data-parallel all-reduce

See merge request ADLR/megatron-lm!989
Move to Draco OCI

See merge request ADLR/megatron-lm!1137
Print number of transformer and embedding parameters separately

See merge request ADLR/megatron-lm!1159
Mcore LLaVA model

See merge request ADLR/megatron-lm!1151
[OMNIML-614] AMMO ptq + TensorRT-LLM export examples for megatron-lm

See merge request ADLR/megatron-lm!1013
Make throughput and memory footprint formulae compatible with arbitrary ffn_hidden_size

See merge request ADLR/megatron-lm!1169
Experimental Yaml configs

See merge request ADLR/megatron-lm!1134
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet