Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evaluate performance #102

Open
Leengit opened this issue Feb 3, 2023 · 2 comments
Open

Evaluate performance #102

Leengit opened this issue Feb 3, 2023 · 2 comments

Comments

@Leengit
Copy link
Collaborator

Leengit commented Feb 3, 2023

In particular, are we leveraging the graph execution optimizations (e.g., parallelization, memory management, GPU usage) of tensorflow and torch or do we need to do more to get that?

@Leengit
Copy link
Collaborator Author

Leengit commented Feb 3, 2023

@cooperlab says: look at TF MultiWorker strategy - https://www.tensorflow.org/tutorials/distribute/multi_worker_with_keras. We can help with this. Key questions are:

  • Can multiple workers run on 1 machine (example suggests so)
  • Can we allow each worker to identify their portion of the "plan" based on their worker index
  • How cumbersome is this - is a convenient wrapper similar to WrappedModel possible for users?

@Leengit
Copy link
Collaborator Author

Leengit commented Feb 21, 2023

Tensorflow does autosharding so we shouldn't have to explicitly shard the tensorflow.Dataset. We could add convenience functions so that the the likes of global_batch_size = num_workers * batch_size_per_worker are satisfied.

If the user has already created a model, and we want to convert or wrap that model so that it is as if the model had been created within a with strategy.scope(): Python block for some distributed strategy, could we do that after the fact? It might work to write the model to disk, and then read it back in within a strategy scope block; I have queried StackOverflow for other possibilities.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant