You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Was experimenting with the mask a bit, and was curious to know whether you ever tested for mathematical equivalence of the masked and unmasked forward steps for Llama?
I tried running unmasked = model(prompt + rejected, mask=None) and then masked = model(prompt + chosen + rejected, mask=block_mask), and then comparing the logits of unmasked and masked[:prompt] + masked[prompt + chosen:]. However, the logits for the rejected part seem to differ in my experiments, at least when using Llama. The results were the same for a 2D bool mask and sdpa.
It's possible this is an issue with my implementation. I'll see if I can reproduce it in your repo.
The text was updated successfully, but these errors were encountered:
Hey, really like this idea!
Was experimenting with the mask a bit, and was curious to know whether you ever tested for mathematical equivalence of the masked and unmasked forward steps for Llama?
I tried running
unmasked = model(prompt + rejected, mask=None)
and thenmasked = model(prompt + chosen + rejected, mask=block_mask)
, and then comparing the logits ofunmasked
andmasked[:prompt] + masked[prompt + chosen:]
. However, the logits for the rejected part seem to differ in my experiments, at least when using Llama. The results were the same for a 2D bool mask and sdpa.It's possible this is an issue with my implementation. I'll see if I can reproduce it in your repo.
The text was updated successfully, but these errors were encountered: