You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# complete this with toolformer.finetune(filtered_results)
92
+
80
93
```
81
94
82
95
The main novelty of the paper is defining a fitness score for the outputs from a transformer instructed to insert API calls. The score is used to filter the sampled outputs for finetuning the transformer to make API calls that decreases perplexity of the text that follows it.
-[ ] Toolformer should eventually calculate all statistics (how many properly sampled, filtered out by different criterias, the distribution of scores as well as how many were rejected) before the final fine-tuning
163
176
-[ ] do end-to-end training in `Toolformer`
164
177
-[x] doing the prompting and bootstrapping the data
165
-
-[] prefiltering of bootstrapped data followed by api calls and then another round of filtering
178
+
-[x] prefiltering of bootstrapped data followed by api calls and then another round of filtering
166
179
-[ ] keep track of all stats
167
180
-[ ] take care of fine-tuning, with the interleaving of datasets + optimizer hyperparams
# deriving the weighting for the original passage is more tricky
474
479
# would need to start counting up from <api> start token location
480
+
# this would also assume that the language model perfectly copied the passage over and that both token ids are aligned except for the inserted API call - but this can be done with the custom filtering functions eventually
475
481
476
482
weight=weight_and_mask_fn(tokens_without_api_response[:, 1:], api_start_token_id) # shift to the left by one since <api> does not exist in the original sequence
assertlen(filtered_data_with_api_calls) >0, 'your model failed to follow instructions and make API calls. please try a better model or do some better prompt engineering'
0 commit comments