Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix the binary model and finetune #3

Open
misaki-taro opened this issue Sep 6, 2023 · 3 comments
Open

Fix the binary model and finetune #3

misaki-taro opened this issue Sep 6, 2023 · 3 comments

Comments

@misaki-taro
Copy link

Hi Jiayu,

I would like to express my appreciation for providing Phavip. However, I have encountered some confusion while working with it. In your documentation, you mention the following: "Thus, we first apply an end-to-end method to train the binary classification model. Then, we fix the parameters in the Transformer encoder and fine-tune a new classifier layer for the multi-class classification model. Binary cross-entropy (BCE) loss and L2 loss are employed for the binary classification and multi-class classification, respectively."

However, when I attempted to retrain the model, I couldn't find the fine-tune mode as described. Could you please assist me in reproducing the results using your code?

Best regards,
Misaki

@misaki-taro misaki-taro changed the title Fix the binary model and fintune Fix the binary model and finetune Sep 6, 2023
@KennthShang
Copy link
Owner

As described, since the model is fixed, you can directly use the given parameters (expect for the final layer) as the pre-trained model.

Best,
Jiayu

@KennthShang
Copy link
Owner

If you do not know how to extract the parameters, you can directly train on a new model without pertaining. The results is almost the same as the increase in the epoch

@misaki-taro
Copy link
Author

Thank you for your advice. I will give it a try at a later time.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants