Skip to content

Q: Padded tokens without an attention_mask?  #10

@Axe--

Description

@Axe--

Hey, I wanted to know how the padded tokens are handled by the attention layers of the model,
considering that repo's implementation of GPT-2 discards attention_mask (HuggingFace).

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions