-
Notifications
You must be signed in to change notification settings - Fork 2.2k
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just a couple questions/comments about organization so far:
- Could "general attention" just be called "attention"?
- Could we put the
T5Attention
in thet5
module? - Can we get rid of
T5AttentionOld
?
Yes to all 3. I'm still debugging a thing with loading pretrained weights, will clean this up in the next commit. This was to get some feedback on the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't have much to say in addition to Dirk's comments. As long as tests pass this LGTM!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just a couple super small suggestions / questions.
We can probably delete allennlp.modules.transformer.self_attention
, right?
Co-authored-by: Pete <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💯
* initial commit * general self attn * fixing bugs, adding tests, adding docs * updating other modules * refactor * bug fix * update changelog * fix shape * fix format * address feedback * small doc fix * Update allennlp/modules/transformer/transformer_stack.py Co-authored-by: Pete <[email protected]> * remove old file Co-authored-by: epwalsh <[email protected]> Co-authored-by: Pete <[email protected]>
Changes proposed in this pull request:
AttentionModule
(naming it so instead of justAttention
to differentiate it fromallennlp.modules.attention.Attention
)SelfAttention
andT5Attention
inherit from this.