Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Getting stuck running training script #12

Open
explor-ar opened this issue Aug 31, 2023 · 1 comment
Open

Getting stuck running training script #12

explor-ar opened this issue Aug 31, 2023 · 1 comment

Comments

@explor-ar
Copy link

Here is what I get in response to running the training code:
C:\Users\AR\Desktop\marlin\MARLIN>python train.py --config config/pretrain/marlin_vit_base.yaml --data_dir C:\Users\AR\Desktop\marlin\MARLIN\trainingData\YouTubeFaces --n_gpus 1 --num_workers 8 --batch_size 16 --epochs 2000 --official_pretrained C:\Users\AR\Desktop\marlin\MARLIN\videomae\checkpoint_vitb.pth _IncompatibleKeys(missing_keys=['encoder.pos_embedding.emb', 'decoder.pos_embedding.emb', 'discriminator.layers.0.linear.weight', 'discriminator.layers.0.linear.bias', 'discriminator.layers.1.linear.weight', 'discriminator.layers.1.linear.bias'], unexpected_keys=[]) GPU available: True (cuda), used: True TPU available: False, using: 0 TPU cores IPU available: False, using: 0 IPUs HPU available: False, using: 0 HPUs Missing logger folder: C:\Users\AR\Desktop\marlin\MARLIN\lightning_logs LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0] Traceback (most recent call last): File "C:\Users\AR\Desktop\marlin\MARLIN\train.py", line 141, in <module> trainer.fit(model, dm) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\trainer\trainer.py", line 696, in fit self._call_and_handle_interrupt( File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\trainer\trainer.py", line 650, in _call_and_handle_interrupt return trainer_fn(*args, **kwargs) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\trainer\trainer.py", line 735, in _fit_impl results = self._run(model, ckpt_path=self.ckpt_path) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\trainer\trainer.py", line 1147, in _run self.strategy.setup(self) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\strategies\single_device.py", line 74, in setup super().setup(trainer) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\strategies\strategy.py", line 153, in setup self.setup_optimizers(trainer) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\strategies\strategy.py", line 141, in setup_optimizers self.optimizers, self.lr_scheduler_configs, self.optimizer_frequencies = _init_optimizers_and_lr_schedulers( File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\core\optimizer.py", line 194, in _init_optimizers_and_lr_schedulers _validate_scheduler_api(lr_scheduler_configs, model) File "C:\Users\AR\AppData\Local\Programs\Python\Python39\lib\site-packages\pytorch_lightning\core\optimizer.py", line 351, in _validate_scheduler_api raise MisconfigurationException( pytorch_lightning.utilities.exceptions.MisconfigurationException: The provided lr scheduler LambdaLRdoesn't follow PyTorch's LRScheduler API. You should override theLightningModule.lr_scheduler_step hook with your own logic if you are using a custom LR scheduler.
I am a beginner with this stuff, so please be forgiving towards my ignorance.

@ControlNet
Copy link
Owner

Please follow the versions in the requirements.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants