Lecture about Swin-Unet
Summary
- Attention-based networks
- Background
- Structured output problems
- Encoder-decoder framework
- Attention Mechanisms
- Image Captioning with Attention Mechanisms
- Transformers
- Background: Machine Translation
- Self-attention
- Scaled dot-product attention
- Multi-head Self Attention module
- Vision domain - DETR
- Background: Object detection & set prediction problem
- Detection Transformer (DETR)
- Vision Transformer - ViT
- Background: Transformers + CNN or RNN
- Vision Transformers - ViT
- Optimizing ViT
- Background: Data-hungry transformers
- Distillation
- Swin Transformer
- Background: Problems using Transformers with images
- Swin Transformer architecture
- Swin Transformer block
- Shifted Window based Self-attention
- Swin-Unet
- Background: How to improve Unets segmentation
- Swin-Unet architecture
- Patch merging and Patch expanding layers