Introduction

Problem with current methods:

SWAT framework:

Understanding Transformer’s Attention