-
Notifications
You must be signed in to change notification settings - Fork 236
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some questions about the implementions and results, particularly related to ROME/MEND/SERAC #417
Comments
As you can see, here are two |
In the ROME hparams configs, I found that you set the default hyper-parameter values of |
I also try to set Additionally, I don't understand the necessity of normalizing u when returning it in |
|
Thanks for your timely reply, and I still have some questions:
|
Another new question: |
|
|
I'm more curious about how to support MEND and SERAC, which require training before editing, in |
You just need to first train the MEND and SERAC module and set them in the hparam files. For training, refer to https://github.com/zjunlp/EasyEdit?tab=readme-ov-file#trainer |
Thank you and I'll check it. |
I have observed that the ZSRE datasets employed for MEND training might exhibit certain discrepancies when compared to the ZSRE test set in the KnowEdit benchmark. Specifically, there are three JSON files involved: zsre_mend_eval.json, zsre_mend_train.json, and zsre_mend_train_10000.json. The zsre_mend_train.json file is notably large(82MB). Given this, it is probable that the training process should utilize the zsre_mend_train_10000.json file instead. This smaller dataset likely corresponds to the training split size of 10,000 instances that you previously mentioned here: https://github.com/zjunlp/EasyEdit?tab=readme-ov-file#dataset But when I try to train on the zsre_mend_train_10000.json, the training process seems never satisfy the early exit demand. So I'll change the setting and try again. |
After I change the layer from 5 to 21, with |
I have tried to firstly train MEDN on zsre_mend_train.json with the default hparams (I just casually pick a ckpt after training step 50000 since the early stop condition seems will not be satisfied until the max training steps 100000, but I don't know why, this is another question), and then use the trained model to conduct single edit on Wikidata_recent, here is my result:
The result seems to be rational but somewhat higher than the reported values. |
Hello!
After thoroughly reviewing the codes related to the implementation of using AdaLoRA and ROME to edit models, I have a few questions:
The text was updated successfully, but these errors were encountered: