-
Notifications
You must be signed in to change notification settings - Fork 60
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Instructions for pruning pre-trained "checkpoint" model #18
Comments
Hi @henrypearce4D ! I've struggled with the same problem as you did but got it working. Hopefully this is helping you to some degree:
No, .ply files are called "pointclouds" and ".pth" files are called checkpoints here. You can start training from either of them, the different arguments are called
Don't use the bash scripts. Take them only as an inspiration for a script you could build yourself. I'll make an example based on the inputs you suggested. Let
I've also added all the default parameters here for better visibility, but it might also work out if you only use the args |
Hi @anton-brandl thanks I will give that a try! This is the section where the example .ply (pointcloud) is referred to as a chpt (checkpoint) and "3D-GS checkpoint" so that was the confusion |
Hi thanks for the great code.
Please could you provide clear instructions for pruning a pre-trained INRIA "checkpoint" model .ply
First off, Checkpoints and Trained models seem to be referenced differently in LightGaussian compared the INRIA code.
Pre-trained INRIA models .ply are being referred to as a checkpoint, this is not a checkpoint in the INRIA code.
INRIA checkpoints are saved with
--checkpoint_iterations 1000
and producechkpnt1000.pth
To run pruning on a pre-trained INRIA "checkpoint" .ply the instructions for LightGaussian say;
Should I add the arguments;
-s path-to-model-folder/
(full path to trained model folder e.g.
-s datasets/big
which contains/point_cloud/iteration_30000/point_cloud.ply
)And
-m path-to-output-folder/
e.g.
bash scripts/run_prune_finetune.sh -s datasets/big -m datasets/small
?In the scripts run_prune_finetune.sh and run_prune__pt_finetune.sh they reference run args for datasets.
In run_prune__pt_finetune.sh a comments says;
# This is an example script to load from ply file.
So should I use this to point directly to .ply file?
bash scripts/run_prune_pt_finetune.sh -datasets/big --start_pointcloud datasets/big/point_cloud/iteration_30000/point_cloud.ply -m datasets/small
I also tired added "big" as an argument to the script.
All these tests failed.
Any help is much appreciated!
The text was updated successfully, but these errors were encountered: