You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, thanks for your excellent job. I've read the paper the reviewed the code. I've encounter some issues as outlined following:
I'd appreciate a detailed explanation of how the "rank classification" is implemented. Could you please provide clarification on the code found at this link?
I'm curious about how the "rank classification" process influences the final results. Is it feasible to employ a direct generation approach, such as generating the label words and matching them against the true answer, as an alternative method?
The text was updated successfully, but these errors were encountered:
The code listed is for training, where we do not use rank classification. Instead, we use the unlikelihood loss that tries to increase the loss for the incorrect labels (i.e. minimizing their probability). This is why there is a negative sign for cross_entropy on line 86 and we mask out the loss for the correct label on line 90.
It is possible to use a direct generation approach and match them against the true answer. Rank classification increases the accuracy of the model since the model only has to choose the correct choice from the list of choice, rather than having to generate the correct choice from the space of all possible outputs.
Hi, thanks for your excellent job. I've read the paper the reviewed the code. I've encounter some issues as outlined following:
I'd appreciate a detailed explanation of how the "rank classification" is implemented. Could you please provide clarification on the code found at this link?
I'm curious about how the "rank classification" process influences the final results. Is it feasible to employ a direct generation approach, such as generating the label words and matching them against the true answer, as an alternative method?
The text was updated successfully, but these errors were encountered: