Skip to content

Commit

Permalink
share idea
Browse files Browse the repository at this point in the history
  • Loading branch information
lucidrains committed Oct 2, 2023
1 parent fd2cc0f commit 78fde6f
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 2 deletions.
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@ Also have a few ideas of my own that I will try and share in this repository, if
- [x] complete prophet net with hierarchical transformer training
- [ ] complete the spec decoding algorithm using trained prophet net transformer

- [ ] for early exit strategy, try randomly summing last cached embedding back to the same model (a la alphafold2 recycling), randomly cropped along sequence length, and train early exit loss this way. see if one can improve the gamma this way
- [ ] dedicate a morning to microoptimizations

## Citations
Expand Down
4 changes: 2 additions & 2 deletions train_prophet.py
Original file line number Diff line number Diff line change
Expand Up @@ -75,8 +75,8 @@ def inner(*args, **kwargs):

prophet = Decoder(
num_tokens = 256,
dim = 128,
depth = 4
dim = 512,
depth = 2
)

model_and_prophet = ModelWithProphetWrapper(
Expand Down

0 comments on commit 78fde6f

Please sign in to comment.