Skip to content

Commit

Permalink
clean up
Browse files Browse the repository at this point in the history
  • Loading branch information
epwalsh committed Feb 24, 2025
1 parent c858268 commit 2d84add
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 2 deletions.
5 changes: 5 additions & 0 deletions src/olmo_core/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -392,6 +392,11 @@ def filter_warnings():
message="Synchronization debug mode is a prototype feature.*",
module="torch.cuda",
)
warnings.filterwarnings(
action="ignore",
category=UserWarning,
message="TORCH_NCCL_AVOID_RECORD_STREAMS=1 has no effect .*",
)
warnings.filterwarnings(
action="ignore",
category=FutureWarning,
Expand Down
3 changes: 1 addition & 2 deletions src/scripts/train/OLMo2-7B-long-context.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
from olmo_core.optim import AdamWConfig, CosWithWarmup, OptimGroupOverride
from olmo_core.train import TrainerConfig
from olmo_core.train.callbacks import CheckpointerCallback, CometCallback, WandBCallback
from olmo_core.train.train_module import ( # TransformerActivationCheckpointingConfig,
from olmo_core.train.train_module import (
TransformerContextParallelConfig,
TransformerDataParallelConfig,
TransformerDataParallelWrappingStrategy,
Expand Down Expand Up @@ -55,7 +55,6 @@ def build_train_module_config(common: CommonComponents) -> TransformerTrainModul
wrapping_strategy=TransformerDataParallelWrappingStrategy.fine_grained,
),
cp_config=TransformerContextParallelConfig(degree=8),
# ac_config=TransformerActivationCheckpointingConfig(),
float8_config=Float8Config(enabled=True),
max_grad_norm=1.0,
scheduler=CosWithWarmup(warmup_steps=2000),
Expand Down

0 comments on commit 2d84add

Please sign in to comment.