- install via
pip install -e .
after cloning and navigating to the repo's main directory - Navigate to any of the task in the tasks folder and run via train.py
decription: Given a sequence of sentences S \ni {s_0, ..., s_n} count the total number of words in sentence s_n, s_{n-2}, s_{n+2}. Sentences are made up of a random sequence of words of varying length.
- For longer sentence lengths the model requires a curriculum learning approach
- Tested for sentences up to 30 words with inputs that contain 100 sentences, the model can solve this although it takes a bit of time
- Here's the weights and biases run for this experiment! https://wandb.ai/wobrob101/word-count/runs/j15svvuj/workspace?nw=nwuserwobrob101
decription:
Given a sequence of sentences S \ni {s_0, ..., s_n}, where a given sentence s_i is made up of a random sequence of words of varying length, for each word in the sentence predict the number of other sentences in S that contain the word.
decription:
Input to the model is a sequence of words W \ni {w_0, ..., w_n}, each word is randomly sampled from a vocabulary, each word in the vocabulary is assigned an integer score in the range {-5, 5}, each word is composed up of tokens, each token is also assigned a score in the range {-5, 5}. The model is then trained to output the culmalitive score of all previous tokens and words aswell as the score for the current token.
(Not implemented yet)
decription:
this task has two settings: reccurent and parallel
Given a sequence of words W \ni {w_0, ..., w_n}, each word is randomly sampled from a vocabulary, each word in the vocabulary is assigned an integer score in the range {-5, 5}, each word is composed up of tokens, each token is also assigned a score in the range {-n, n}. The model is trained to predict a final score for each token. This is produced by taking the score for the word at token index t and adding to the score of the token at index t + token_scores[token_index], this is performed starting at token index 0 and ending at token index n.
Same as above except all of the addition operations are performed at once, rather than starting at token zero and ending at token n.