You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I am wondering if it is possible to intervene a model which contains both some layers of encoders and some layers of decoders. And I would like to assign different intervention locations for these two kinds of transformer blocks. I found that in the pyvene library, it just supports one intervention_locations argument. If it is possible to realize things like only prefix (7) for encoders and f7+l7 for decoders without modifying the original library? Thank you!
The text was updated successfully, but these errors were encountered:
Also note, decoder interventions are harder to implement correctly. You need to intervene on a variety length of input IDs, which complicates collating and padding. But theoretically there is nothing that prevents implementing encoder-decoder interventions. Pyvene/Pyreft is just a model intervening library that can do anything. Checkout my PR for an example.
Hi, I am wondering if it is possible to intervene a model which contains both some layers of encoders and some layers of decoders. And I would like to assign different intervention locations for these two kinds of transformer blocks. I found that in the pyvene library, it just supports one intervention_locations argument. If it is possible to realize things like only prefix (7) for encoders and f7+l7 for decoders without modifying the original library? Thank you!
The text was updated successfully, but these errors were encountered: