ET: Re-Thinking Self-Attention for Transformer Models on GPUsPublished in SC 21 (to appear), 2021Share on Twitter Facebook LinkedIn Previous Next