Swin Transformer is an important work among all the attempts to reduce the computational complexity of Transformers while maintaining its excellent performance in computer vision. Window-based patch self-attention can use the local connectivity of the image features, and the shifted window-based patch self-attention enables the communication of information between different patches in the entire image scope. Through in-depth research on the effects of different sizes of shifted windows on the patch information communication efficiency, this article proposes a Dual-Scale Transformer with double-sized shifted window attention method.
View Article and Find Full Text PDFOptical flow is widely used in medical image processing, such as image registration, segmentation, 3D reconstruction, and temporal super-resolution. However, high-precision optical flow training datasets for medical images are challenging to produce. The current optical flow estimation models trained on these non-medical datasets, such as KITTI, Sintel, and FlyingChairs are unsuitable for medical images.
View Article and Find Full Text PDF