Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ttnn implementation for Swin_S model #8590

Open
saichandax opened this issue May 17, 2024 · 1 comment
Open

ttnn implementation for Swin_S model #8590

saichandax opened this issue May 17, 2024 · 1 comment
Assignees

Comments

@saichandax
Copy link
Contributor

saichandax commented May 17, 2024

reference model is from: https://github.com/pytorch/vision/blob/main/torchvision/models/swin_transformer.py

@HariniMohan0102
Copy link
Contributor

HariniMohan0102 commented May 17, 2024

To implement Swin_s transformer model, model is segregated into submodules where each module is dependent on the other. The sequence of submodules are

  1. ShiftedWindowAttention
  2. SwinTransformerBlock
  3. PatchMerging
  4. SwinTransformer

Currently working on ShiftedWindowAttention submodule.
Task Updates regarding ShiftedWindowAttention:

  • Reference implementation completed.
  • Compared reference with torch, PCC=1.0
  • ttnn implementation is in progress.
  • Submodule to be tested further.

For your reference, added torch and referenc graphs of ShiftedWindowAttention submodule.
Torch graph:
torch_graph

Reference graph:
ref_graph

saichandax added a commit that referenced this issue May 29, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants