Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ pinned: false
|
|
11 |
|
12 |
*By: [Hila Chefer](https://hila-chefer.github.io) and [Sayak Paul](https://sayak.dev)*
|
13 |
|
14 |
-
*Website: [
|
15 |
|
16 |
*Abstract: The attention mechanism has revolutionized deep learning research across many disciplines, starting from NLP and expanding to vision, speech, and more. Different from other mechanisms, the elegant and general attention mechanism is easily adaptable and eliminates modality-specific inductive biases. As attention become increasingly popular, it is crucial to develop tools that allow researchers to understand and explain the inner workings of the mechanism to facilitate better and more responsible use of it. This tutorial focuses on understanding and interpreting attention in the vision, and in the multi-modal settings combining text and vision. We present state-of-the-art research on representation probing, interpretability, and attention-based semantic guidance, alongside hands-on demos to facilitate interactivity. Additionally, we discuss open questions arising from recent works and future research directions.*
|
17 |
|
|
|
11 |
|
12 |
*By: [Hila Chefer](https://hila-chefer.github.io) and [Sayak Paul](https://sayak.dev)*
|
13 |
|
14 |
+
*Website: [all-things-vits.github.io/atv/](https://all-things-vits.github.io/atv/)*
|
15 |
|
16 |
*Abstract: The attention mechanism has revolutionized deep learning research across many disciplines, starting from NLP and expanding to vision, speech, and more. Different from other mechanisms, the elegant and general attention mechanism is easily adaptable and eliminates modality-specific inductive biases. As attention become increasingly popular, it is crucial to develop tools that allow researchers to understand and explain the inner workings of the mechanism to facilitate better and more responsible use of it. This tutorial focuses on understanding and interpreting attention in the vision, and in the multi-modal settings combining text and vision. We present state-of-the-art research on representation probing, interpretability, and attention-based semantic guidance, alongside hands-on demos to facilitate interactivity. Additionally, we discuss open questions arising from recent works and future research directions.*
|
17 |
|